Interactions Between Computational Intelligence and Mathematics Compress (1)
Interactions Between Computational Intelligence and Mathematics Compress (1)
László T. Kóczy
Jesús Medina Editors
Interactions
Between
Computational
Intelligence and
Mathematics
Studies in Computational Intelligence
Volume 758
Series editor
Janusz Kacprzyk, Polish Academy of Sciences, Warsaw, Poland
e-mail: [email protected]
The series “Studies in Computational Intelligence” (SCI) publishes new develop-
ments and advances in the various areas of computational intelligence—quickly and
with a high quality. The intent is to cover the theory, applications, and design
methods of computational intelligence, as embedded in the fields of engineering,
computer science, physics and life sciences, as well as the methodologies behind
them. The series contains monographs, lecture notes and edited volumes in
computational intelligence spanning the areas of neural networks, connectionist
systems, genetic algorithms, evolutionary computation, artificial intelligence,
cellular automata, self-organizing systems, soft computing, fuzzy systems, and
hybrid intelligent systems. Of particular value to both the contributors and the
readership are the short publication timeframe and the world-wide distribution,
which enable both wide and rapid dissemination of research output.
Editors
Interactions Between
Computational Intelligence
and Mathematics
123
Editors
László T. Kóczy Jesús Medina
Faculty of Engineering Sciences Departamento de Matemáticas,
Széchenyi István University Facultad de Ciencias
Gyõr Universidad de Cádiz
Hungary Puerto Real, Cádiz
Spain
and
v
vi Preface
Weighted Mean (WM) and the Weighted Ordered Average (OWA). A new family,
merging the advantages of both previously mentioned classes, the Weighted OWA
Operators (WOWA), offers a combination of advantages of the abovementioned
two methods. The paper describes in detail the procedure of filtering the SAR
images, with all three approaches. For learning the parameters, the authors use a
Genetic Algorithm (GA), the classical evolutionary approach. As the GA is not one
of the most efficient evolutionary methods, it is no surprise that the authors con-
clude that there is a strong dependence of the quality of the results on the number of
generations. The final result points toward all proposed filters being useful, or even
at the best—depending on the type of image and the circumstances of application.
The directions of future research are indicated at the end.
The third paper proposes a novel combination of wavelet transform and fuzzy
rule interpolation for evaluating very big data obtained from the telecommunication
field. The engineering problem is the pre-evaluation of copper wire pairs from the
point of view of transmission speed in SHDSL transmission networks. In the
present best practice, the insertion loss must be measured at a large number of
different frequencies for every wire pair. Based on several 100,000s of real mea-
surement results, it is proposed that the combination of the Haar and Daubechies-4
wavelets combined with the Stabilized KH rule interpolation leads to a much better
prediction of the transmission quality than any method known so far in the liter-
ature. The conclusion suggests that this new method might be useful in a much
wider field of big data analysis, as well.
The fourth paper introduces a logic called the Molecular Interaction Logic,
which semantically characterizes the Molecular Interaction Maps (MIM) and,
moreover, makes it possible to apply deductive and abductive reasoning on MIMs
in order to find inconsistencies, answer queries, and infer important properties about
those networks. This logic can be applied to different metabolic networks, as the
one given by the cancer, which can sometimes appear in a cell as a result of some
pathology in a metabolic pathway.
The fifth paper is based on another logic. In this case in the general fuzzy logic
programming called multi-adjoint logic programming. This framework was intro-
duced by Medina et al. in 2001 as a general framework in which the minimal
mathematical requirements are only considered in order to ensure the main prop-
erties given in the diverse usual logic programming frameworks, such as in
possibilistic logic programming, monotonic and residuated logic programming,
fuzzy logic programming, etc. Since its introduction, this framework has been
widely developed from the theoretical and applied aspects by diverse authors. In
this paper, the authors perform experiments which have shown the benefits of using
a new c-unfolding transformation which reuses some variants of program
transformation techniques based on unfolding, which have been largely exploited
in the pure functional—not fuzzy—setting. Specifically, this new c-unfolding
transformation has been applied to fuzzy connectives, and the authors have shown
how to improve the efficiency of the proper unfolding process by reusing the very
well-known concept of dependency graph. Furthermore, the paper includes a cost
analysis and discussions on practical aspects.
Preface vii
The last two papers are in the area of the Formal Concept Analysis (FCA). This
framework arose as a mathematical theory for qualitative data analysis and has
become an interesting research topic both on its mathematical foundations and on
its multiple applications. Specifically, the sixth paper provides an overview of
different generalizations of formal concept analysis based on fuzzy sets. First of all,
a common platform for early fuzzy approaches is included. Then, different recently
fuzzy extensions have been recalled, such as the generalized extension given by
Krajči, the multi-adjoint concept lattices introduced by Medina et al., heterogeneous
extension studied by diverse research groups exemplified by those of Krajči,
Medina et al., Pócs, Popescu, using heterogeneous one-sided extension given by
Butka and Pócs, and the higher order extension presented by Krídlo et al. The paper
also discusses connections between the related approaches.
The seventh paper is related to knowledge extraction from databases using rules,
which is a compact tool in the representation of the knowledge. Based on concepts
from lattice theory, this paper has introduced a new kind of attribute implications
considering the fuzzy notions of support and confidence. The authors have also
studied different properties of the particular case in which the set of attributes are
intensions and, finally, they have also included an application to clustering for size
reduction of concept lattices.
We would like to express our gratitude to all the authors for their interesting,
novel, and inspiring contributions. Peer-reviewers also deserve our deep appreci-
ation because their deep and valuable remarks and suggestions have considerably
improved the contributions.
And last but not least, we wish to thank Dr. Tom Ditzinger, Dr. Leontina di
Cecco, and Mr. Holger Schaepe for their dedication and help to implement and
finish this large publication project on time maintaining the highest publication
standards.
ix
Page Rank Versus Katz: Is the Centrality
Algorithm Choice Relevant to Measure
User Influence in Twitter?
1 Introduction
Nowadays, there are 288 million active users on Twitter and more than 500 million
tweets are produced per day [17]. Through short messages, users can post about their
This work was supported by national funds through Fundação para a Ciência e a
Tecnologia (FCT) under project PTDC/IVC-ESCT/4919/2012 and funds with reference
UID/CEC/50021/2013.
feelings, important events and talk amongst each other. Twitter has become so much
of a force to be reckoned with, that anybody from major brands and institutions, to
celebrities and political figures use it to further assert their position and make their
voice heard. The impact of Twitter on the Arab Spring [6] and how it beat the all news
media to the announcement of Michael Jackson’s death [15], are just a few examples
of Twitter’s role in society. When big events occur, it is common for users to post
about it in such fashion, that it becomes a trending topic, all the while being unaware
from where it stemmed or who made it relevant. The question we wish to answer is:
“Which users were important in disseminating and discussing a given topic?”
Much like real life, some users carry more influence and authority than others.
Determining user relevance is vital to help determine trend setters [16]. The user’s
relevance must take into account not only global metrics that include the user’s level
of activity within the social network, but also his impact in a given topic [18]. Empir-
ically speaking, an influential person can be described as someone with the ability
to change the opinion of many, in order to reflect his own. While [13] supports this
statement, claiming that “a minority of users, called influentials, excel in persuading
others”, more modern approaches [4] seem to emphasize the importance of inter-
personal relationships amongst ordinary users, reinforcing that people make choices
based on the opinions of their peers.
In [2], three measures of influence were taken into account: “in-degree is the
number of people who follow a user; re-tweets mean the number of times others
forward a user’s tweet; and mentions mean the number of times others mention a
user’s name”. It concluded that while in-degree measure is useful to identify users
who get a lot of attention, it “is not related to other important notions of influence such
as engaging audience”. Instead “it is more influential to have an active audience who
re-tweets or mentions the user”. In [8], the conclusion was made that within Twitter,
“news outlets, regardless of follower count, influence large amounts of followers to
republish their content to other users”, while “celebrities with higher follower totals
foster more conversation than provide retweetable content”. The authors in [12]
created a framework named “InfluenceTracker”, that rates the impact of a Twitter
account taking into consideration an Influence Metric, based on the ratio between
the number of followers of a user and the users it follows, and the amount of recent
activity of a given account. Much like [2], it also shows “that the number of followers
a user has, is not sufficient to guarantee the maximum diffusion of information (…)
because, these followers should not only be active Twitter users, but also have impact
on the network”.
In this paper, we analyze how two well known network analysis algorithms,
PageRank and Katz, affect the computation of mention-based user influence in Twit-
ter. Although these two methods have previously been compared [11] and found
to have been equivalent, we show that the same conclusion does not apply in the
context of social networks, and that PageRank is indeed more adequate. We base
our conclusions on a real world case study of the 2011 London Riots, since it was
an important social event where Twitter users were said to have played a role in its
origin and dissemination.
Page Rank Versus Katz: Is the Centrality Algorithm Choice Relevant to … 3
3.1 PageRank
Arguably the most well known centrality algorithm is PageRank [9]. It is one of
Google’s methods to its search engine and it was created as way for computing a
ranking for every web page based on the graph of the web uses. In this algorithm,
web pages are nodes, while back-links form the edges of the graph (Fig. 1). It is
defined by Eq. 1 as P R(vi ) of a page vi .
1−d P R(v j )
P Rvi = +d (1)
N v ∈M(v )
L(v j )
j i
It can be intuitively said about Eq. 1, that a page has high rank if the sum of
the ranks of its back-links is high. In it, v j is the sum ranges over all pages that
has a link to vi , L(v j ) is the number of outgoing links from v j , N is the number
of documents/nodes in the collection and d is the damping factor. The PageRank
is considered to be a random walk model, because the weight of a page vi is “the
probability that a random walker (which continues to follow arbitrary links to move
from page to page) will be at vi at any given time. The damping factor corresponds to
4 H. Rosa et al.
the probability of the random walk to jump to an arbitrary page, rather than to follow
a link, on the Web. It is required to reduce the effects on the PageRank computation
of loops and dangling links in the Web” [11]. Dangling links are “simply links that
point to any page with no outgoing links (…) they affect the model because it is
not clear where their weight should be distributed” [9]. The true value that Google
uses for damping factor is unknown, but it has become common to use d = 0.85 in
the literature. A lower value of d implies that the graph’s structure is less respected,
therefore making the “walker” more random and less strict.
3.2 Katz
It was shown in [11] that “Katz status index may be considered a more general
form of PageRank because in can be modified, within a reasonable range, to be
equivalent to PageRank” and that under a “relaxed definition of equivalence (…)
PageRank and Katz status index is practically equivalent to each other” as long as
the number of outgoing links from any vertex is the same throughout the graph,
which is very unlikely for graph modeled from a social network. On the other hand,
“it is also possible to modify PageRank to become completely equivalent to Katz
status index”, however, in that case, “the modified PageRank is no long a random
work model because it can no longer be modeled from a probabilistic standpoint”
[11].
Page Rank Versus Katz: Is the Centrality Algorithm Choice Relevant to … 5
4 Dataset
In order to test the network analysis methods presented above, a database from the
London Riots in 2011 [3] was used. The London Riots of 2011 was an event that
took place between the 6th and 11th August 2011, where thousands of people rioted
in several boroughs of London with the resulting chaos generated looting, arson, and
mass deployment of police. Although Twitter was said to be a communication tool
for rioting groups to organize themselves, there is little evidence that it was used
to promote illegal activities at the time, though it was useful for spreading word
about subsequent events. According to [5], Twitter played a big role spreading the
news about what was happening and “was a valuable tool for mobilizing support for
the post-riot clean-up and for organizing specific clean-up activities”. Therefore it
constitutes a prime data sample to study how users exert influence in social networks,
when confronted with such a high stakes event.
The Guardian Newspaper made public a list of tweets from 200 influential twitter
users, which contains 17,795 riot related tweets and an overall dataset of 1,132,938
tweets. Using a Topic Detection algorithm [1], we obtained an additional 25,757
unhastagged tweets about the London Riots. It consists of a Twitter Topic Fuzzy
Fingerprint algorithm [14] that provides a weighted rank of keywords for each topic
in order to identify a smaller subset of tweets within scope. This method has proven
to achieve better results than other well known classifiers in the context of detecting
Topics within Twitter, while also being faster in execution. The sum of posting and
mentioned users is 13,765 (vertices) and it has 19,993 different user mentions (edges),
edges
achieving a network connectivity ratio of ver tices
= 1.46.
In this section, we compare the results of ranking the most influential users using Page
Rank, Katz and a mentions based baseline. We proceed by performing an empirical
analysis of the users in order to ascertain their degree of influence and their position
in the ranks. The graphs and ranking were calculated using Graph-Tool [10].
Table 1 shows how both network analysis algorithms behave while highlighting
the rank differences (shown by the arrows in the last column). A “Mentions rank”
is used as a base line. Figure 2 provides a visual tool to the graph, as provided by
PageRank.
There is an obvious relation between the number of mentions and the ranking pro-
vided by the application of both algorithms: the highest ranked users in either Katz and
PageRank, are some of the most mentioned users in our dataset. In fact, the relation is
more clear between Katz and the baseline Mentions based ranking: Table 1 shows that
the rank in both approaches is always either identical (@guardian, @skynewsbreak,
@gmpolice, etc…) or at most separated by two positions (@richardpbacon is ranked
27th based on mentions, and 29th based on Katz). In order to determine the relation
6
Table 1 London riots top 20 most influential users according to page rank, and comparison with Katz
User Mentions PageRank Katz
# Rank Score Rank Score Rank
@guardian 160 2 0.0002854 1 0.022157 2
@skynewsbreak 178 1 0.0002512 2 0.023479 1
@gmpolice 122 4 0.0002128 3 0.019009 4
@riotcleanup 107 6 0.0001767 4 0.017992 6
@prodnose 67 14 0.0001761 5 0.014022 15
@metpoliceuk 116 5 0.0001494 6 0.018709 5
@marcreeves 69 11 0.0001476 7 0.014195 12
@piersmorgan 78 8 0.0001465 8 0.014959 9
@scdsoundsystem 69 12 0.0001442 9 0.014190 13
@subedited 70 10 0.0001337 10 0.014278 11
@youtube 48 20 0.0001257 11 0.012424 20
@bbcnews 94 7 0.0001256 12 0.016426 8
@mattkmoore 62 15 0.0001237 13 0.013614 16
@richardpbacon 40 27 0.0001218 14 0.011771 29
@lbc973 34 35 0.0001150 15 0.011432 34
@skynews 74 9 0.0001113 16 0.014638 10
@bengoldacre 61 17 0.0001055 17 0.013526 17
@bbcnewsnight 68 13 0.0000988 18 0.014123 14
@tom_watson 44 21 0.0000968 19 0.012107 22
@paullewis 129 3 0.0000954 20 0.019602 3
...
@juliangbell 61 16 0.0000275 188 0.0166597 7
The arrows indicate most relevant rank differences
H. Rosa et al.
Page Rank Versus Katz: Is the Centrality Algorithm Choice Relevant to … 7
Fig. 2 User influence page rank graph - larger circles indicate larger user influence
between PageRank and “Mentions Rank”, the Spearman correlation was calculated
having achieved a value of ρ = 0.9372, which means they are heavily correlated.
However, when limiting this calculation to the top 20, it changed to ρ = 0.5535,
which implies that for the top users, just looking at the number of mentions, is not
enough to determine influence.
An empirical analysis also shows that both Page Rank and Katz largely agree
upon the ranking of most users, namely on the top two users: (i) @guardian, Twitter
account of the world famous newspaper “The Guardian”; (ii) @skynewsbreak, Twit-
ter account of the news team at Sky News TV channel. This outcome agrees with [8]
previous statement, that, “news outlets, regardless of follower count, influence large
amounts of followers to republish their content to other users” and can be justified by
the incredibly high London Riots news coverage. Other users seem to fit the profile,
namely @gmpoliceq, @bbcnews and @skynews. Most of the other users are either
political figures, political commentators or jornalists (@marcreeves, @piersmorgan,
and @mattkmoore).
However, when looking more closely at Page Rank versus Katz rankings, it is
also possible to realize some notorious differences: Katz’s third and seventh top
ranked users are not in PageRank’s top users. The reasons behind these differences
in the ranking positions should be thoroughly analyzed since they could highlight
the strengths and weaknesses of each algorithm in what concerns their capability to
8 H. Rosa et al.
express user influence in social networks. The two cases end up being different and
should be treated separately: (i) @paullewis, ranked 3rd by Katz shows up at 20th
according to PageRank; (ii) @juliangbell, ranked 7th by Katz shows up at 188th
according to PageRank.
The reason behind @paullewis high placement in the Katz rank is the number
of mentions. As said previously, Katz is a generalization of a back-link counting
method, which means the more back-links/mentions a user has, the higher it will be
on the ranking. This user has 129 mentions, but PageRank penalizes it, because it is
mentioned by least important users, which means a less sum weight is being trans-
fered to it in the iterative process. This logic also applies to users @bbcnewsnight,
@skynews and @bbcnews. Additionally, @paullewis is also an active mentioning
user, having mentioned other users a total of 14 tweets, while @skynewsbreak and
@guardian have mentioned none. As a consequence, Paul Lewis transfers its influ-
ence across the network while the others simply harvest it. There are several users
that drop in ranking from PageRank to Katz for the very same reason. Users such as
@prodnose, @marcreeves and @youtube do not have enough mentions for Katz to
rank them higher.
User @juliangbell, despite mentioned often (61 times), is down on the PageRank
because of indirect gloating, i.e., he retweets tweets that are mentioning himself:
“@LabourLocalGov #Ealing Riot Mtg: @juliangbell speech http://t.co/3BNW0q6”
was posted by @juliangbell himself. The user is posting somebody else’s re-tweet
of one of his tweets. As a consequence a link/edge was created from @juliangbell
to @LabourLocalGov, but also from @juliangbell to himself, since his username is
mentioned in his own tweet. Julian Bell is a political figure, making it acceptable
that he would have a role in discussing the London Riots, but the self congratulatory
behavior of re-tweeting other people’s mentions of himself, is contradictory with
the idea of disseminating the topic across the network. While Katz is not able to
detect this effect, PageRank automatically corrects it, which is why, contrary to
what is mentioned in previous works [11], it is our comprehension that Katz is not
equivalent to PageRank in the task of detecting user relevance in social networks
such as Twitter.
6 Conclusions
With this study, we have shown that in the context of user influence in Twitter,
PageRank and Katz are not equal in performance, thus disproving previous claims.
PageRank has proved a more robust solution to identify influential users in discussing
and spreading a given relevant topic, specially when considering how it deals with
indirect gloating, an item Katz fails to penalize.
Page Rank Versus Katz: Is the Centrality Algorithm Choice Relevant to … 9
References
1. Carvalho, J.P., Pedro, V., Batista, F.: Towards intelligent mining of public social networks’
influence in society. In: IFSA World Congress and NAFIPS Annual Meeting (IFSA/NAFIPS),
pp. 478 – 483. Edmonton, Canada (June 2013)
2. Cha, M., Haddadi, H., Benevenuto, F., Gummadi, K.P.: Measuring user influence in twitter: the
million follower fallacy. In: In ICWSM ’10: Proceedings of International AAAI Conference
on Weblogs and Social (2010)
3. Crockett, K.S.R.: Twitter riot dataset (tw-short) (2011)
4. Domingos, P., Richardson, M.: Mining the network value of customers. In: Proceedings of the
Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Min-
ing, pp. 57–66. KDD ’01, ACM, New York, USA (2001). http://doi.acm.org/10.1145/502512.
502525
5. Guardian, T.: http://www.theguardian.com/uk/2011/dec/07/twitter-riots-how-news-spread
6. Huang, C.: Facebook and twitter key to arab spring uprisings: report. http://www.thenational.
ae/news/uae-news/facebook-and-twitter-key-to-arab-spring-uprisings-report (June 2011).
Accessed 02 May 2014
7. Katz, L.: A new status index derived from sociometric analysis. Psychometrika 18(1), 39–43
(March 1953). http://ideas.repec.org/a/spr/psycho/v18y1953i1p39-43.html
8. Leavitt, A., Burchard, E., Fisher, D., Gilbert, S.: The Influentials: New Approaches for Ana-
lyzing Influence on Twitter (2009)
9. Page, L., Brin, S., Motwani, R., Winograd, T.: The Pagerank Citation Ranking: Bringing Order
to the Web (1999)
10. Peixoto, T.: https://about.twitter.com/company
11. Phuoc, N.Q., Kim, S.R., Lee, H.K., Kim, H.: Pagerank vs. katz status index, a theoretical
approach. In: Proceedings of the 2009 Fourth International Conference on Computer Sciences
and Convergence Information Technology, pp. 1276–1279. ICCIT ’09, IEEE Computer Society,
Washington, DC, USA (2009). http://dx.doi.org/10.1109/ICCIT.2009.272
12. Razis, G., Anagnostopoulos, I.: Influencetracker: Rating the Impact of a Twitter Account.
CoRR (2014). arXiv:1404.5239
13. Rogers, E.M.: Diffusion of Innovations (1962)
14. Rosa, H., Batista, F., Carvalho, J.P.: Twitter topic fuzzy fingerprints. In: WCCI2014, FUZZ-
IEEE, 2014 IEEE World Congress on Computational Intelligence, International Conference
on Fuzzy Systems, pp. 776–783. IEEE Xplorer, Beijing, China (July 2014)
15. Sankaranarayanan, J., Samet, H., Teitler, B.E., Lieberman, M.D., Sperling, J.: Twitterstand:
News in tweets. In: Proceedings of the 17th ACM SIGSPATIAL International Conference on
Advances in Geographic Information Systems, pp. 42–51. GIS ’09, ACM, New York, NY, USA
(2009). http://doi.acm.org/10.1145/1653771.1653781
16. Tinati, R., Carr, L., Hall, W., Bentwood, J.: Identifying communicator roles in twitter. In:
Proceedings of the 21st International Conference Companion on World Wide Web, pp. 1161–
1168. WWW ’12 Companion, ACM, New York, NY, USA (2012). http://doi.acm.org/10.1145/
2187980.2188256
17. Twitter: https://about.twitter.com/company
18. Weng, J., Lim, E.P., Jiang, J., He, Q.: Twitterrank: finding topic-sensitive influential twitterers.
In: Proceedings of the Third ACM International Conference on Web Search and Data Mining,
pp. 261–270. WSDM ’10, ACM, New York, NY, USA (2010). http://doi.acm.org/10.1145/
1718487.1718520
Weighted Means Based Filters for SAR
Imagery
Abstract We address parameter learning for three families of filters for SAR
imagery, based on WM, OWA and WOWA families of aggregation operators. The
values in the weight vector associated to a WM filter correspond to the same posi-
tions in the input, whereas those in OWA filters consider the ordered positions of
the input. WOWA operators make use of both vectors to weight an input data vector.
Here we use Genetic Algorithms to learn the weight vectors for OWA, WM and
WOWA filters and assess their use in reducing speckle in SAR imagery. We present
an application using simulated images derived from a real-world scene and compare
our results with those issued by a set of filters from the literature.
1 Introduction
Synthetic Aperture Radar (SAR) sensors are not as adversely affected by atmospheric
conditions and the presence of clouds as optical sensors [13]. Moreover, they can
be used at any time of day or night. The visual quality of SAR images is, however,
degraded by sudden variations in image intensity with a salt and pepper pattern, due
to the existence of a great amount of speckle, a multiplicative non-Gaussian noise,
proportional to the intensity of the received signal [17]. Speckle in SAR image
hampers the interpretation and analysis of this kind of image, as well as reduces the
effectiveness of some image processing tasks such as segmentation and classification.
For this reason, in order to process a SAR images, it is usually recommended to apply
a filter before the segmentation or classification processes, even though the image
quality with respect to some features may decrease, such as signal to noise ratio,
spatial resolution, among others.
Filters for SAR imagery can be classified according to whether they take into
account a model for speckle. A class of important representative of model-independent
filters is the Ordered Statistical Filters (OSF) [2], based on order statistics [18].
WOWA filters for SAR imagery are learned with the adopted GA. Section 5 presents
an experiment in SAR imagery and Sect. 6 finally brings the conclusion.
Given a window in an image, a filter simply substitutes the value of its central pixel by
a function of the values of the pixels in the window. Two of the most basic filters use
the arithmetic mean and the median as filtering function. In SAR imagery, the mean
filter tends to reduce the speckle but it also tends to indiscriminately blur the image
[14]. The median filter, on the other hand, reduces erratic variations by eliminating
the lowest and highest pixel values [21].
Most filters employ the convolution operation. Given an image I , whose pixels
take values in R, a m × m window around the central pixel (x, y) in I , and a matrix
of coefficients γ : {−m, ..., 0, ..., m}2 → R, the result of convolution for (x, y) in
14 L. Torres et al.
In Order Statistics Filters [2], the result of filtering for a given pixel is the linear
combination of the ordered values of the pixels in the window around that pixel.
They belong to the larger class of non-linear filters based on order statistics [18],
being an application of L-estimators. An OSF is thus obtained when a convolution
filter is applied on the ordered statistic of the pixel values in a window.
Some examples of model-independent filters are the directional means filters, in
which only pixels in one of the twelve regions of the six orthogonal directions are
considered (diagonals, rows and columns) [21] and the local region filters (see [22]),
in which the window is divided in eight regions based on angular position, and the
central pixel is replaced by the mean value of the subregion presenting the lowest
variance.
The adoption of a model to the noise leads to more complex filters. One of such
filters is the so-called Lee filter, in which speckle reduction is based on multiplicative
noise model using the minimum mean-square error (MMSE) criterion [10, 11]. The
Refined Lee filter [12], here called R-Lee filter, is an improved version of the Lee
filter, and uses a methodology for selecting neighboring pixels with similar scattering
characteristics.
Buades et al. [3] proposed a model-dependentt methodology that is well-suited for
decreasing additive Gaussian noise, called Nonlocal Means (NL-means), which uses
similarities between patches as the weights of a mean filter. The SDNLM (Stochastic
Distances and Nonlocal Means) filter [27] is an adaptive nonlinear extension of the
NL-means algorithm filter, in which overlapping samples are compared based on
stochastic distances between distributions, and the p-values resulting from such
comparisons are used to build the weights of an adaptive linear filter.
Assessing the performance of image filters is very hard [34]. Two important indices
to measure the quality of filtered images are NMSE and SSIM, described in the
following. Index NMSE (Normalized Mean Square Error) is a general purpose error
measure, widely used in image processing (see [1]). Let r be the perfect information
data and s an approximation of r ; NMSE is calculated as:
n
j=1 (r j − s j )
2
NMSE = n 2
, (1)
j=1 r j
Weighted Means Based Filters for SAR Imagery 15
where r j and s j refer to values in r and s at the same coordinates (the position of a
given pixel in the case of images). NMSE always yield positive values, and the lower
its value, the better is the approximation considered to be.
Index SSIM (Structural SIMilarity) measures the similarity between two scalar-
valued images and can be viewed as a quality measure of one of them, when the other
image is regarded as of perfect quality [35]. It is an improved version of the universal
image quality index proposed proposed by [33]. This index takes into account three
factors: (i) correlation between edges; (ii) brightness distortion; and (iii) distortion
contrast. Let r and s be the perfect information and its approximation, respectively;
SSIM is calculated as
Cov(r, s)+α1 2r s +α2 σr
2 σs +α3
SSIM(r, s) = × 2 2 × 2 , (2)
σr
σs +α1 r +s +α2 σr +σs2 +α3
In the following, we first give the definitions of WM, OWA and WOWA families of
operators. We then describe the filters obtained using these operators.
Let w be a weighting vector of dimension n (w= [w1 w2 ... wn ]), such that:
– (i) wi ∈ [0, 1];
– (ii) Σi wi = 1.
16 L. Torres et al.
where {σ (1), ..., σ (n)} is a permutation of {1, ..., n}, such that aσ (i−1) ≥ aσ (i) for
all i = {2,..., n} (i.e., aσ (i) is the i-th largest element in {a1 , ..., an }).
Some well-known OWA operators are the mean, min, max and median, which are
obtained with OWA vectors wmean , wmin , wmax , and wmed , respectively. For instance,
taking n = 3, we have: wmean = [1/3, 1/3, 1/3], wmin = [0, 0, 1], wmax = [1, 0, 0],
and wmed = [0, 1, 0].
The measures of or ness, and andness [36], associated with a given vector w are
defined as follows:
or ness(w) = n−1
1
Σi (n − i)wi
andness(w) = 1 − or ness(w)
Functions or ness and andness describe an adjustment of the levels of “or” and
“and”, respectively, in the aggregation of a set of values.
wowa
Let p and w be weighting vectors as given above. A mapping fw, p : R n → R is a
Weighted Ordered Weighted Average (WOWA) operator of dimension n, associated
to p and w, if [24]:
wowa
f w, p,φ (a1 , ..., an ) = Σi ωi × aσ (i) , (5)
where {σ (1), ..., σ (n)} is a permutation of {1, ..., n}, for all i = {2,..., n}, such that
aσ (i−1) ≥ aσ (i) . Weight ωi is defined as
We now present a simple example to illustrate the use of OWA, WM and WOWA
operators. Let a1 = 10, a2 = 20, and a3 = 0. We thus have σ (1) = 2, σ (2) = 1, and
σ (3) = 3. Therefore aσ (1) = 20, aσ (2) = 10, and aσ (3) = 0.
Let w = [0.5 0.3 0.2] and p = [1/6 2/3 1/6]. We thus obtain
– f wowa (10, 20, 0) = 0.5 × 20 + 0.3 × 10 + 0.2 × 0 = 13,
– f pwm (10, 20, 0) = 1/6 × 10 + 4/6 × 20 + 1/6 × 0 = 15.
Let φl be a linear by parts function, given by
where {(xi , yi )i=1,n } is a set of predetermined points. Using w to obtain the points
{(xi , yi )i=1,n } as required in Eq. 6, and then applying them in Eq. 8, we obtain φl (0) =
0, φl (1/3) = 0.5, φl (2/3) = 0.8 and φl (1) = 1.
Taking σ and p above, we have pσ (1) = 2/3, pσ (2) = 1/6, and pσ (3) = 1/6. We
then obtain Pσ (1) = 2/3, Pσ (2) = 5/6, and Pσ (3) = 1, as well as φl (Pσ (1)) = 0.8,
φl (Pσ (2)) = 0.9, and φl (Pσ (3)) = 1. Therefore ω1 = 0.8, ω2 = 0.1, and ω3 = 0.1.
Thus
wowa
– f w, p,φ (10, 20, 0) = 0.8 × 20 + 0.1 × 10 + 0.1 × 0 = 17.
OWA (respec. WM) filters (see [28, 29]) are obtained by applying OWA (respec.
WM) weight vectors in the values inside a sliding window over a given image. Both
WM and OWA filters are convolution filters, where coefficients γ are positive and
add up to 1, with the coefficients in OWA being applied in the order statistic of the
data, which makes them OSFs [7].
In the present paper, we introduce WOWA filters, based on a combination of WM
wowa
and OWA filters. Given an image I , the application of a WOWA filter Fw, p,φ derives
wowa
a filtered image Iw, p,φ , as described below. OWA and WM filters Fw and F pwm can
owa
wowa
Procedure Fw, p,φ (I )
6. Make the result become the value for position (x, y) in the filtered image:
wowa wowa
Iw, p,φ (x, y) = f w, p,φ (a1 , ..., an ).
In the following, we describe our framework to learn the weights for OWA, WM
and WOWA filters for SAR imagery. We first describe Genetic Algorithms, the
optimization algorithm adopted in this work, and then how it is adapted to SAR
imagery specificities.
Genetic Algorithms (GA), first proposed in [9] (see also [6]), combine Mendel’s
ideas about the codification of life in genes, with Darwin’s ideas on the survival of
the fittest (natural selection). They are search algorithms that evolve populations of
candidate solutions, according to a fitness function that assesses the quality of these
solutions to solve the problem at hand.
A candidate solution c ∈ C = {c1 , . . . , ck } consists of a set of parameters to a
function sol, that models the problem at hand. Each c can be thought of as a genotype
(chromosome) and sol(c) as its corresponding phenotype. A fitness function f it
evaluates the candidate solutions; f it (sol(c)) should be proportional to the capacity
of c ∈ C in solving the problem at hand.
At each GA iteration, three processes (selection, crossover and mutation) take
place, generating a new population C . The selection process is such that the fittest
candidates in C have a higher probability of being selected for reproduction. This
process is usually performed by means of a roulette (the larger the fitness of an
individual, the larger its share in the roulette wheel) or a set of tournaments (at each
tournament, a set of individuals are chosen at random from the population and the
winner is selected for reproduction). Different forms of elitism can also be used,
by forcing the best candidates to be remain in the new population and/or to have a
stronger influence on the creation of C . The reproduction process, called crossover,
creates two new candidate solutions by mixing the genotypes of two selected parent
candidate solutions. In the mutation process, all new candidate solutions can suffer
changes, according to a (usually small) probability, called the mutation rate, here
denoted as ρ.
Weighted Means Based Filters for SAR Imagery 19
The first initial population is usually obtained at random, but in many applications,
the use of a selected set of chromosomes may lead to better results. The stop criterion
is usually a fixed number of iterations. The combination of selection, crossover and
mutation provide GAs with a good equilibrium between exploration and exploitation
of the search space.
In our work, each chromosome consists of weight vectors and each position in
the chromosome contains a real number (a weight), that sum up to 1 altogether.
The crossover operator adopted here consists in a linear combination of the parent
chromosomes values. Given two parents c1 and c2 , and random number α ∈ [0, 1],
we generate two sons c12 and c21 , where ∀i ∈ {1, ..., n}, c12 [i] = α × c1 [i] + (1 −
α) × c2 [i] and c21 [i] = (1 − α) × c1 [i] + α × c2 [i]. The values in c12 and c21 are
then normalized to guarantee that the weights sum up to 1.
In previous works, we tested a few strategies for mutation (see e.g. [28, 29]),
and in the following we describe the ones that gave the best results, called A and
B. In these strategies, the mutation rate is not applied on each position but on the
chromosome as a whole. If a chromosome is selected for mutation, we randomly
select the main position 1 ≤ q ≤ n in the chromosome to be changed, considering a
uniform distribution. Then these mutation strategies differ as follows:
– A: The value in position q is multiplied by mutation rate ρ. The difference is
divided by n − 1 and added to each of the remaining positions. Note that the
larger is ρ, the larger is the change of the value in position q and the smaller is the
change in the other positions in the vector.
– B: The value in q is increased with the value of its neighbour, considering the
chromosome as a ring, and the neighbour receives value 0. When the neighbour
to the right (respec. left) is considered, the strategy is named Br (respec. Bl).
In our previous works with OWA filters in SAR imagery, we verified that mutation
operator A usually outperformed its B counterparts.
Given a SAR image on a given area, we need to have the means to assess the quality
of filters applied on it. As done in previous works (see [28, 29]), here we adopt the
following framework:
– samples from classes of interest are taken from a SAR image,
– for each class, the parameters of its associated distribution are estimated,
– using a phantom image, in which the regions are associated to the classes in the
original image, a set of simulated images is created at random using the class
distributions,
– the set of simulated images is partitioned in two sets, one for training and one for
testing, and
20 L. Torres et al.
– the best weight vectors found by the GA on the training set is used on the test set
for evaluation.
Considering a m × m standard window, WM and OWA filters require chromo-
somes with n = m 2 positions to encode weight vectors p and w, respectively. As for
WOWA filters, when p and w are learned at the same time as here, each chromosome
has n = 2m 2 positions.
In [29], we investigated the use of two strategies to learn OWA weight vectors for
k polarizations, k > 1:
– S1: k GAs are run independently, one for each polarization, with weight vectors
containing n positions.
– S2: a single GA is run, with the weight vector containing k × n; when the weight
vector is learned, it is split in k parts, resulting in a weight vector of n positions
for each polarization.
The second strategy requires a more complex dealing with the GAs than the first one:
even though selection, mutation and crossover remain basically the same, we have
to ensure consistency, with each of the k parts of the chromosome adding up to 1.
In our works, we verified that S2 produced better results for OWA filters in terms of
quality of results and computational efficiency.
5 Experiments
(a) SAR image fragment [23] (b) Samples (ROI) (c) Phantom fragment [20]
Fig. 1 SAR L-band false color composition, using HH (red), HV (green) and VV (blue) polariza-
tions of the area of interest, ROIs and phantom used in the experiments
all filters. Here we report the use of OWA, WM and WOWA filters, considering the
following parametrizations:
– selection type: roulette,
– number of generations: 30 and 120,
– population size: 36 elements,
– mutation rates: 0.2 and 0.8,
– seeds for random numbers: 2 and 271.
For each experiment, we performed a 5-fold cross-validation, using 40 images
for training and 10 for testing in each fold. The elements in the initial population in
each experiment were chosen at random. As fitness function for each fold in each
parametrization, we took the means of the quality of the resulting filtered images,
according to index NMSE (see Sect. 2.2).
Tables 1, 2 and 3 bring the results obtained for WM, OWA and WOWA filters,
respectively, with parameters by GAs, according to NMSE, considering 5 × 5 win-
dows. In the tables, we denoted the experiments as E/ξ /ψ, where ξ ∈ {30, 120} and
ψ ∈ {0.2, 0.8}. Therefore, an experiment run with 30 generations and mutation rate
0.2 is denoted by E/30/0.2. We mark in bold the best results obtained in each seed.
The best aggregated results between the two seeds are marked with an asterisk (“∗ ”).
Table 1 NMSE mean (truncated) using WM on 50 images, with GA filters learned using 5 fold
cross-validation, 5 × 5 windows, and mean NMSE of training images as fitness function
Seed 2 Seed 271
HH HV VV Agg HH HV VV Agg
E/30/.2 0.0547 0.0561 0.0570 0.0559 0.0540 0.0541 0.0555 0.0545
E/30/.8 0.0527 0.0535 0.0545 0.0536 0.0522 0.0532 0.0544 0.0533
E/120/.2 0.0499 0.0508 0.0518 0.0508∗ 0.0500 0.0508 0.0518 0.0509
E/120/.8 0.0504 0.0511 0.0523 0.0513 0.0503 0.0512 0.0520 0.0512
22 L. Torres et al.
Table 2 NMSE mean (truncated) using OWA on 50 images, with GA filters learned using 5 fold
cross-validation, 5 × 5 windows, 36 elements in each population and mean NMSE of training
images as fitness function
Seed 2 Seed 271
HH HV VV Agg HH HV VV Agg
E/30/.2 0.0515 0.0523 0.0536 0.0525 0.0518 0.0523 0.0537 0.0526
E/30/.8 0.0516 0.0523 0.0536 0.0525 0.0517 0.0524 0.0536 0.0526
E/120/.2 0.0486 0.0494 0.0505 0.0495∗ 0.0486 0.0494 0.0506 0.0495∗
E/120/.8 0.0486 0.0495 0.0505 0.0495∗ 0.0486 0.0494 0.0506 0.0495∗
Table 3 NMSE mean (truncated) using WOWA on 50 images, with GA filters learned using 5 fold
cross-validation, 5 × 5 windows, and mean NMSE of training images as fitness function
Seed 2 Seed 271
HH HV VV Agg HH HV VV Agg
E/30/.2 0.0515 0.0524 0.0538 0.0526 0.0515 0.0522 0.0536 0.0524
E/30/.8 0.0515 0.0523 0.0536 0.0525 0.0515 0.0523 0.0535 0.0524
E/120/.2 0.0486 0.0511 0.0512 0.0503 0.0493 0.0503 0.0506 0.0501∗
E/120/.8 0.0499 0.0497 0.0511 0.0502 0.0494 0.0502 0.0513 0.0503
In Table 1, we see that the mean NMSE of all results range between 0.0499 and
0.0570. The best aggregated result was obtained using 120 generations, seed 2 and
mutation rate 0.2. We see in this table is that a large number of generations indeed
increases the performance of WM filters.
In Table 2, we see that the mean NMSE of all results are more similar in the various
parametrizations than in the case of WM, ranging between 0.0486 and 0.0537. The
best aggregated result was obtained using 120 generations, seeds 2 and 271 and
mutation rates 0.2 and 0.8. We see that also here, the most important factor on the
results is the number of generations, with the seeds and mutation rates producing
basically the same results, considering the individual polarizations as well as the
aggregated results.
In Table 3, we see that the mean NMSE of all results are very close to those
obtained by OWA filters, ranging between 0.0486 and 0.0538. The best aggregated
result was obtained using 120 generations, seed 271 and mutation rate 0.2. Again
here, the large number of generations produced the best results.
Comparing together the results of Tables 1, 2 and 3, we see that WM filters have
been outperformed by both OWA and WOWA filters in all experiments. The results of
OWA and WOWA are very similar; WOWA filters produced better results than OWA
with a smaller number of generations, however, with a large number of generations,
OWA produces the best overall results. In a nutshell, WOWA outperformed WM in
our experiments but did not outperform OWA with our choice for function φ.
Weighted Means Based Filters for SAR Imagery 23
Table 4 brings the results obtained by the best aggregated GA-learned filters,
considering the same type of window and the same number of folds. In Table 4, we
also report the results for SDNLM and R-Lee filters, with the best parametrizations
chosen after a few experiments, using 5 × 5 filtering window for both filters, with
3 × 3 patches, and significance level of 5% for for SDNLM, and with ENL = 1 for
R-Lee. The best WM, OWA and WOWA filters for 5 × 5 windows are taken from
Tables 1, 2 and 3, with ties solved considering non displayed decimal cases. For both
5 × 5 and 3 × 3 windows, the number of elements in the population is 36, and the
best number of generations and best mutation rate are 120 and 0.2, respectively.
Considering both types of windows, the seed value for OWA and WM is 2, whereas
for WOWA the seed value is 271.
In Table 4, we see that the WOWA, OWA and WM filters, with weight vectors
learned with a GA with 5 × 5 windows, outperformed all other filters considered
here with respect to NMSE (used by the fitness function). In particular, OWA filters
produced the best results, both in the individual polarizations as well as when they
are aggregated. In what regards SSIM, the best performances were obtained with
SDNLM and the mean filter. In the remaining of this section, we only discuss NMSE
results.
Table 4 NMSE and SSIM mean (truncated) on 50 images, with GA filters learned using 5 fold
cross-validation, 5 × 5 windows, and mean NMSE of training images as fitness function
NMSE SSIM
HH HV VV Agg HH HV VV Agg
Unfiltered 0.997 0.074 1.025 1.008 0.068 0.084 0.053 0.068
SDNLM 0.078 0.074 0.119 0.091 0.166 0.177 0.126 0.157
(5×5)
R-Lee (5×5) 0.079 0.075 0.125 0.093 0.155 0.167 0.124 0.149
OWA (5×5) 0.048 0.049 0.050 0.049 0.147 0.158 0.147 0.151
WOWA 0.049 0.050 0.050 0.050 0.146 0.157 0.147 0.150
(5×5)
WM (5×5) 0.049 0.050 0.051 0.050 0.148 0.159 0.148 0.152
Mean (5×5) 0.064 0.063 0.124 0.084 0.166 0.178 0.124 0.156
Median (5×5) 0.153 0.153 0.203 0.170 0.123 0.131 0.099 0.118
Min (5×5) 0.927 0.928 0.934 0.930 0.001 0.001 0.000 0.000
Max (5×5) 9.754 9.868 9.590 9.737 0.024 0.026 0.019 0.023
OWA (3×3) 0.106 0.107 0.108 0.107 0.133 0.145 0.130 0.136
WOWA 0.111 0.110 0.110 0.110 0.132 0.145 0.130 0.135
(3×3)
WM (3×3) 0.117 0.118 0.118 0.117 0.133 0.146 0.130 0.136
Mean (3×3) 0.115 0.116 0.116 0.116 0.137 0.150 0.133 0.140
Median (3×3) 0.189 0.190 0.190 0.190 0.107 0.116 0.104 0.109
Min (3×3) 0.807 0.807 0.808 0.807 0.007 0.007 0.007 0.007
Max (3×3) 4.969 5.025 4.971 4.988 0.039 0.046 0.037 0.040
24 L. Torres et al.
In [28], we have verified that filters learned with 3 × 3 windows did not fare so
well when compared with the model-based filters SDNLM and R-Lee, which use
5 × 5 windows, considering a GA configuration of 30 generations at most. That
indicated that 5 × 5 windows are more appropriate 3 × 3 ones for our filters. Here,
we see that with 120 generations, the model-based filters still outperform our filters
with 3 × 3 windows but are outperformed when 5 × 5 windows are used. We see
that even a very large increase in the number of generations in the GA is not capable
of overcoming the negative impact that small windows cause on our filters.
Figure 2 brings an unfiltered synthetic image and the filtered images obtained from
it using some of the methods considered here. We note that WM, OWA and WOWA
(a) Unfiltered (0.9991, 0.0699) (b) OWA (0.0525, 0.1626) (c) WOWA (0.0526, 0.1622)
(d) WM (0.0533, 0.1637) (e) Mean (0.0863, 0.1504) (f) SDNLM (0.0912, 0.1579)
(g) R-Lee (0.0953, 0.1509) (h) Min (0.9270, 0.0010) (i) Max (9.6841, 0.0234)
Fig. 2 False color composition of results from the same simulated images, considering methods
using 5 × 5 windows, for HH (red), HV (green) and VV (blue) polarizations, with mean NMSE
and SSIM from the polarizations inside parentheses
Weighted Means Based Filters for SAR Imagery 25
in Fig. 2 are visually superior to those obtained with both complex filters; SDNLM
produces a blurred image and the Lee filter yields a pixelated image.
Figure 3 brings a square fragment of the false composition of the original image,
with the ROIs from which the samples for each class were extracted (Fig. 3a),1 the
images obtained with OWA, WOWA and WM filters (Fig. 3b–d), all using as weights
the best parametrization learned in fold 3, and the model-independent SDNLM and
R-Lee filters (Fig. 3e, f, resp.).
Figure 3b, c, d show the effect of the OWA, WOWA and WM on windows of size
5 × 5 over the whole image. Albeit the noise reduction is evident, it is also clear that
the blurring introduced eliminates useful information as, for instance, curvilinear
details in the river area.
Figure 3f is the result of applying the R-Lee filter; although presenting a good
performance, some details in the edges are eliminated, being worse than all the
previous filters in this aspect. Figure 3e present the result of smoothing the original
data set with the SDNLM filter at the level significance α = 10%. The noise effect is
alleviated, e.g. the graininess is reduced specially in the planting and over the river
areas, but fine details are more preserved than when the R-Lee filter is employed.
Another important point is that land structures within the river are enhanced, and
their appearance is maintained.
Figures 4 and 5 respectively illustrate the weight vectors w and p found in our
best experiments. We see that the weights found are not distributed homogeneously
(specially in what regards VV), but the the highest values for HH and HV are around
the median. The orness found for vectors w for HH, HV and VV are respectively
0.494, 0.496 and 0.511, which is coherent with the visual inspection. The weights
in vectors p, all polarizations considered, varied between 0.009 and 0.067. We see
that HH and HV produce a more homogeneous distribution of weights than VV.
Although there are large weight values in the borders, we see that the central pixel in
all polarizations have a high weight value. We also conducted a set of experiments
(not shown here) in which we learned a smaller number of weights that were repeated,
according to the relative positions to the central pixel for p, and to the median for w,
but the results were inferior to those obtained here.
The GA was run on a machine with the following specifications: Intel i7, CPU 2.60
GHz, RAM with 16 GB, Windows 10, Fortran with Force 2.0 compiler. Considering 5
folds, with 10 images in each fold, 36 elements in the population, and 30 generations,
the GA processing for all our filters (WM, OWA and WOWA) took approximately
30 min for each polarization for 3 × 3 windows and approximately 1 h for 5 × 5
windows. When the number of generations is multiplied by 4 (120), the computation
times also increase 4 times, in all cases.
1 The available image with ROIs use the Pauli composition, that takes the complex part of the images
into account. For this reason, the colours are different from the filtered images, that use the same
composition used in Fig. 1a, which only takes the intensity images into account.
26 L. Torres et al.
(a)Unfiltered(ROI) (b)OWA
(c)WOWA (d)WM
(e)SDNLM (f)R-Lee
Fig. 3 False color composition of the original SAR image unfiltered, with ROIs, and images
obtained with a set of filters
Weighted Means Based Filters for SAR Imagery 27
Fig. 5 Weight vectors p found in the best experiment (values multiplied by 102 with a single digit
displayed)
In this paper, we addressed OWA, WM and WOWA filters, based on OWA, WM and
WOWA operators (see [25, 36]), respectively. We compared their ability to reduce
speckle in SAR imagery, considering simulated images from three polarizations (HH,
HV and VV) and two types of windows (3 × 3 and 5 × 5). The weight vectors of
the three filters were learned with Genetic Algorithms, for both a small and a large
number of generations.
Our experiments have shown that the number of generations used in the G.A. has
a strong impact on the quality of the results. They have also shown that the size of
the windows is however the most crucial factor for obtaining good results.
We compared the results against a set of filters from the literature, including
model-independent filters WM and OWA (which also had their parameters learned)
and two model-dependent ones, SDNLM [27], a recent parametrized family of filters
(here with, parameters chosen after trial-and-error), and the well-known Refined Lee
filter [12]. OWA filters, followed by WOWA and WM filters, with parameters learned
using 5 × 5 windows outperformed all other filters, according to indice NMSE, used
in the learning process.
The quality of filters depend to what one wants to extract from the filtered images.
Visual inspection indicates that the sole use of one index to assess the quality of
filters may not be enough for some applications. We are currently investigating the
use multi-objective optimization to produce filters that take into account not only the
28 L. Torres et al.
error, measured by NMSE, but the preservation of the structures in the image, using
a measure based on edge-detection.
This work and previous ones show that automatic learning of parameters can be
crucial for parametrized families of filters. For instance, the filters obtained using the
min, max, median and arithmetic mean operators are all particular cases of the OWA
and WOWA operators and fared significantly worse than their learned counterparts.
To be fair, the results obtained by OWA and WOWA filters should also be compared
to model-dependent filters whose parametrizations are also learned, but this is out of
the scope of this work.
As future research, we intend to use other alternatives to model function φ, used
by the WOWA operators, instead of the linear by parts function adopted here. We
also intend to study the influence of initial populations in the GA, and the use of
other types of windows, such as the ones based on non-standard neighbourhoods
(see, e.g. [5]). Last but not least, we intend to use the whole information contained in
a POLSAR image, taking into account also the complex components, and not only
the real components, as we have done so far.
Acknowledgements The authors are indebted to Vicenç Torra and to Lluis Godo for discussions
on WOWA operators, and to Benicio Carvalho for providing computational resources.
References
1. Baxter, R., Seibert, M.: Synthetic aperture radar image coding. MIT Lincoln Lab. J. 11(2),
121–158 (1998)
2. Bovik, A.C., Huang, T.S., Munson, D.C.: A generalization of median filtering using linear
combinations of order statistics. IEEE Trans. ASSP 31(6), 1342–1349 (2005)
3. Buades, A., Coll, B., Morel, J.M.: A review of image denoising algorithms, with a new one.
Multiscale Model. Simul. 4(2), 490–530 (2005)
4. Carlsson, C., Fullér, R.: Fuzzy reasoning in decision making and optimization. In: Studies in
Fuzziness and Soft Computing Series. Springer (2002)
5. Fu, X., You, H., Fu, K.: A statistical approach to detect edges in SAR images based on square
successive difference of averages. IEEE GRSL 9(6), 1094–1098 (2012)
6. Goldberg, D.E.: Genetic Algorithms in Search, Optimization, and Machine Learning. Addison-
Wesley (1989)
7. Grabisch, M., Schmitt, M.: Mathematical morphology, order filters and fuzzy logic. In: Pro-
ceedings of FuzzIEEE 1995, Yokoham, pp. 2103–2108 vol. 4 (1995)
8. Herrera, F.: Genetic fuzzy systems: status, critical considerations and future directions. Int. J.
Comput. Intell. Res. 1(1), 59–67 (2005)
9. Holland, J.H.: Adaptation in Natural and Artificial Systems. University of Michigan Press,
USA (1975)
10. Lee, J.-S., Grunes, M.R., de Grandi, G.: Polarimetric SAR speckle filtering and its implication
for classification. IEEE Trans. GRS 37(5), 2363–2373 (1999)
11. Lee, J.-S., Grunes, M.R., Mango, S.A.: Speckle reduction in multipolarization, multifrequency
SAR imagery. IEEE Trans. GRS 29(4), 535–544 (1991)
12. Lee, J.-S., Grunes, M.R., Schuler, D.L., Pottier, E., Ferro-Famil, L.: Scattering-model-based
speckle filtering of polarimetric SAR data. IEEE Trans. GRS 44(1), 176–187 (2006)
13. Lee, J.-S., Pottier, E.: Polarimetric radar imaging: from basics to applications. In: Optical
Science and Engineering. Taylor and Francis, UK (2009)
Weighted Means Based Filters for SAR Imagery 29
14. Mascarenhas, N.: An overview of speckle noise filtering in sar images. In: 1st Latin-American
Seminar on Radar Remote Sensing—Image Processing Techniques, pp. 71–79 (1997)
15. Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial
domain. IEEE Trans. Image Process. 21(12), 4695–4708 (2012)
16. Moschetti, E., Palacio, M.G., Picco, M., Bustos, O.H., Frery, A.C.: On the use of Lee’s protocol
for speckle-reducing techniques. Lat. Am. Appl. Res. 36(2), 115–121 (2006)
17. Mott, H.: Remote Sensing with Polarimetric Radar. Wiley, USA (2006)
18. Pitas, I., Venetsanopoulos, A.N.: Nonlinear Digital Filters: Principles and Applications.
Springer, Berlin (2013)
19. Richards, J.A.: Remote Sens. Imaging Radar. Springer, Signals and Communication Technol-
ogy Series (2009)
20. Saldanha, M.F.S.: Um segmentador multinível para imagens SAR polarimétricas baseado na
distribuição Wishart. Ph.D. Thesis, INPE, Brazil (2013)
21. SantAnna, S.J.S., Mascarenhas, N.: Comparação do desempenho de filtros redutores de
“speckle” . In: VIII SBSR, pp. 871–877 (1996)
22. Sheng, Y., Xia, Z.G.: A comprehensive evaluation of filters for radar speckle suppression. In:
IGARSS ’96, vol. 3, pp. 1559 – 1561 (1996)
23. Silva, W.B., Freitas, C.C., Sant’Anna, S.J.S., Frery, A.C.: Classification of segments in PolSAR
imagery by minimum stochastic distances between Wishart distributions. IEEE J-STARS 6(3),
1263–1273 (2013)
24. Torra, V.: The weighted OWA operator. Int. J. Intell. Syst. 12(2), 153–166 (1997)
25. Torra, V.: On some relationships between the WOWA operator and the Choquet integral. In:
Proceedings of IPMU’98, Paris, France (1998)
26. Torra, V., Narukawa, Y.: Modeling Decisions: Information Fusion and Aggregation Operators.
Springer, Berlin (2007)
27. Torres, L., Sant’Anna, S.J.S., Freitas, C.C., Frery, A.C.: Speckle reduction in polarimetric SAR
imagery with stochastic distances and nonlocal means. Pattern Recogn. 47(1), 141–157 (2014)
28. Torres, L., Becceneri, J.C., Freitas, C.C., Sant’Anna, S.J.S., Sandri, S.: OWA filters for SAR
imagery. In: Proceedings of LA-CCI’15, Curitiba, Brazil, pp. 1–6 (2015)
29. Torres, L., Becceneri, J.C., Freitas, C.C., Sant’Anna, S.J.S., Sandri, S.: Learning OWA filters
parameters for SAR imagery with multiple polarizations. In: Yang, X.-S., Papa, J.P. (eds.) Bio-
Inspired Computation and Applications in Image Processing, pp. 269–284. Elsevier, Nether-
lands (2016)
30. Torres, L., Becceneri, J.C., Freitas, C.C., Sant’Anna, S.J.S., Sandri, S.: WOWA image filters.
In: Proceedings of CBSF’16, Campinas, Brazil (2016)
31. Torres, L., Becceneri, J.C., Freitas, C.C., Sant’Anna, S.J.S., Sandri, S.: Comparing OWA and
WOWA filters in mean SAR images. In: Proceedings of SBSR’17, S.J.Campos, Brazil (2017)
32. Ulaby, F.T., Elachi, C.: Radar Polarimetry for Geoscience Applications. Artech House, USA
(1990)
33. Wang, Z., Bovik, A.C.: A universal image quality index. IEEE SPL 9(3), 81–84 (2002)
34. Wang, Z., Bovik, A.C., Lu, L.: Why is image quality assessment so difficult? In: IEEE ICASSP,
vol. 4, pp. 3313–3316, Orlando (2002)
35. Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error
visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–612 (2004)
36. Yager, R.R.: On ordered weighted averaging aggregation operators in multi-criteria decision
making. IEEE Trans. Syst. Man Cybern. 18, 183–190 (1988)
On Combination of Wavelet
Transformation and Stabilized KH
Interpolation for Fuzzy Inferences Based
on High Dimensional Sampled Functions
1 Introduction
Due to the great number of input values, making inference on phenomena which can
be described by large-sized vectors is difficult and expensive. In order to construct
efficient inference systems, simplification of the input space is needed. This sim-
plification makes the process of the inference easier, however, it unavoidably rises
the system’s level of uncertainty and inaccuracy. During our previous research on
performance prediction of physical links of telecommunications access networks,
we had to encounter such problems in two ways.
Due to the limited calculation capacity not all the measured values can be used as
the bases of the inference. Drastically lowering the number of the measured frequency
dependent input values caused an inaccuracy in the final results. Later, this type of
sparseness will be referred to as vertical sparseness.
As it is not possible to measure all possible data, the finite teaching sample set
will naturally result in sparse rule bases: there will be points, which will be outside
of all the supports of the antecedent sets of the corresponding dimensions. Later, this
behaviour will be referred to as horizontal sparseness.
In Sect. 2 the primary technical problem underlying the research on performance
prediction is briefly reviewed, the first version of the inference method, its test results
and the horizontal and vertical sparseness derived from the simplification of the input
space are also shown. In Sect. 3 wavelet transformation and fuzzy rule interpolation
as the algorithmic techniques applied in a combined way for handling the problems
of simplification are described, and in Sect. 4 we present the test results of the new
approach based on these techniques.
insertion loss series were clustered into 5 groups according to the measured bit rates.
This grouping is similar to the practice of telecommunication service providers when
offering packages of DSL services.
Based on insertion loss values at 6 well-selected characteristic frequencies, fuzzy
rule bases were created to predict the maximal available SHDSL bit rate of sym-
metrical wire pairs. The result of this prediction is the label of one of the bit rate
groups. Two types of rule bases were created. One of them was constructed directly
from the measured values. It consists of five six dimensional rules, in which trian-
gular antecedent and consequent fuzzy sets are used. The input dimensions are the
insertion loss values that can be measured at the 6 characteristic frequencies. Each
rule can be unambiguously assigned to one of the output states. Using the measured
and clustered data as teaching samples, another type of rule bases were created by
bacterial evolutionary algorithm [4], resulting in a rule base with trapezoidal fuzzy
sets and ten rules. This rule base has the same 6 dimensions: the 6 characteristic
frequencies [5]. Examples of the rule antecedents from each type of rule base can be
seen in Fig. 1. In the figure, the upper diagram belongs to the triangular rule base, and
the lower one to the rule base constructed by bacterial evolutionary algorithm. The
upper rule is obviously sparse. Even though seemingly there are no gaps between the
trapezoidal fuzzy sets in the rule base made by the bacterial evolutionary algorithm,
this rule base can be also considered as a sparse or incomplete one, because insertion
loss values can be measured outside of the supports of all fuzzy sets, as well.
The above two rule bases were tested by the measurements of more than 60 wire
pairs in operating access networks and there were no relevant differences between
their respective results. In most of the cases, where all measured values belonged
to insertion loss areas covered by antecedent sets, the predictions were successful.
Only 13 lines out of 65 could be evaluated, and the predictions were correct in case
of 12 lines form this 13.
[6] D6
0
0 20 40 60 80
Insertion Loss [dB]
1
0,5
0
0 20 40 60 80
Insertion Loss [dB]
34 F. Lilik et al.
1,538%
80%
However correct the results of the successful predictions were, the high proportion
(80%) of the lines (Fig. 2) where no results were produced due to the gaps in the rule
base (no overlap with any antecedent), is not acceptable in practice. This phenomenon
was caused by two reasons.
The reason for unsuccessful evaluations, where no valid results were produced,
was the insufficient knowledge of the observed physical phenomenon. Obtaining
only a limited amount of valid data during the measurement, only sparse rule bases
could be constructed. Measurement results of wire pairs to be tested contained values
also outside of the supports of the antecedent fuzzy sets, hence, in these cases there
could be no valid conclusion calculated. This deficiency of the method is considered
as a “horizontal sparseness”.
The reason for valid, but incorrect prediction was investigated as well. This was
caused by the drastically lowered dimensionality of the input space (values only at
6 points out of 150 were considered). Even though, generally, the insertion loss
measured at the selected characteristic frequencies led to correct predictions, in
several, infrequently occurring cases, measured values at these 6 frequency points
could have such deviations which could bias the result of the prediction. Using
only several variables from the whole input space can be considered as “vertical
sparseness” of the model of the reality.
In the next section a new approach will be proposed, suitable for the elimination
of both mentioned problems.
Vertical sparseness of the rule bases was derived from the partial usage of the possible
input data. It was needed in order to decrease the dimensionality of the applied fuzzy
inference system, however, a large amount of information of the measured insertion
loss functions was wasted. Finding a method which keeps the simplicity of the fuzzy
On Combination of Wavelet Transformation and Stabilized … 35
system and the information of the used insertion loss functions was needed. As
wavelet transformation is efficient in reducing the size of any continuous or discrete
functions down to a required level, it seemed to be successfully applicable in the
problem.
Horizontal sparseness of the fuzzy system, namely the sparseness of the rule
bases, can be handled by the techniques of fuzzy rule interpolation. Stabilized KH
interpolation fits continuous and mathematically stable functions to all α-cuts of the
membership functions in the rules, which can treat the observations in the gaps and
out of the domains of the rules, too (in this way performing also extrapolation).
Basics of wavelet transformation and stabilized KH interpolation are briefly
overviewed in the followings.
with the window function w, which usually is either compactly supported or has
very quickly decaying tails towards the positive and negative infinity. This provides
a short sample of the function f to be transformed. However, if the window function
is combined with the transforming exponential, a short wave
It can be seen from the above formula, that all the wavelets ψb,a (x) are generated
from one mother wavelet ψ(x) by dilating and shifting, thus whereas in case of
the windowed Fourier transform, the grid distance and the size of the transforming
function remains the same at all resolutions, only its shape changes, for wavelet
transforms, the shape of the functions remains the same and both the grid and the
width of the function shrinks as the resolution (i.e., the frequency or spatial frequency)
increases. This property is one of the driving forces behind the success of wavelet
analysis, as usually the high frequency terms are localized in smaller spatial domains
than the slowly varying parts, thus changing the window size with the frequency is
usually very effective.
Wavelet analysis can be carried out by a series of filter pairs. There is a high-pass
and a low-pass filter in all of the pairs, as it can be seen in Fig. 3, the high-pass ones
(after a downsampling) giving the wavelet components and the low-pass ones being
transformed further. The downsampling steps in each of the branches in Fig. 3 take
only every second element of the result and neglect the points in between. In each of
such steps the frequency limit of the low-pass distribution is at around the half of the
highest frequency of the incoming distribution. It can be seen easily, that low pass
outputs give a coarse-grained, or averaged behaviour of the distribution, whereas the
wavelet terms provide the fine-scale details.
The total number of the elements in the resulting vectors ci and di is almost the
same as that of the original vector ci , only a slight increase might arise due to the
size of the filters. The size of the filters Ns can be different for the different wavelet
types, however it is typically less than 20—in image processing usually less than
Fig. 3 One filter pair of the discrete wavelet transform. After the high pass and low pass convolu-
tional filters and the downsamplings the transformed vectors ci and di arise, their size is about half
of the size of the original ci
On Combination of Wavelet Transformation and Stabilized … 37
for the high-pass coefficients. The transforming function φba is a so called scaling
function, and ψba a wavelet, as previously. From these coefficients the function f
can be approximated at resolution level 2−A as
∞
∞
A−1
f A (x) = cb0 φb0 (x) + gba φba (x), (5)
−∞ a=0 −∞
One step of this synthesis procedure of the wavelet transform is summarized in Fig. 4,
where the synthesis filters are after an upsampling step that introduces zeros between
the elements of the incoming vectors. Usually, more of these steps are following each
other with introducing finer and finer resolution details.
In data analysis—also in our case—the starting point is a sampled function and
the end result is the lowest resolution level low pass vector and the high pass vectors.
Our starting vector is a series of insertion loss values measured at consecutive fre-
quency points, and the resulting vectors will give information about the large-scale
behavior of the insertion loss vs. frequency function. In the following considerations
Fig. 4 Synthesis steps of the wavelet transform. The synthesis filters are plotted with yellow
rectangles and the upsampling step with a rectangle including an upright arrow
38 F. Lilik et al.
Daubechies’s [7] wavelet and scaling function sets are used with 2 and 4 nonzero
filter coefficients. In case of the 2-element filter, the averaging process is without
overlaps with the neighboring frequency domains, whereas, in case of the 4-element
filter, the domains of the weighted averages overlap.
Transformations of the starting sampled insertion loss functions were carried out
until only 10 and 5 vector elements remained. As the power spectral density of the
transmission is larger in the lower frequency domain, we have merged the results of
the two vectors so that the points would meet our previously selected characteristic
frequency points.
and
2n
k
1
sup{Biα }
dαU (A∗ , Ai )
sup{Bα∗ } =
i=1
2n k , (7)
1
i=1
dαU (A∗ , Ai )
where i denotes the number of the rules, k the number of the dimensions (variables),
A∗ the observation, Ai the antecedent sets in rule i, dαL (A∗ , Ai ) and dαU (A∗ , Ai ) the
lower and upper bounds of the distance between the α-cuts of observation and the
antecedents, and B ∗ stands for the corresponding fuzzy conclusion [20]. In practice
it is efficient to calculate the values of Bα∗ for α = 0 and 1.
On Combination of Wavelet Transformation and Stabilized … 39
In order to avoid the problems reviewed in Sect. 1, the techniques of Sects. 3.1 and
3.2 were used.
First, the wavelet transformed version of the insertion loss values used in rule base
construction were calculated. Daubechies-2 (Haar) [21] and Daubechies-4 wavelets
were used and the transformations were performed down to 5 points resolution as
described in the previous section. Figure 5 shows the original and the Haar wavelet
transformed insertion loss values as an example. As a matter of course, wavelet
transformation results in discrete values, however, to make the corresponding points
visible, they are graphically linked in the figure.
New rule bases were created by the Wavelet transformed insertion loss series. In
this pattern, the rule base based on Daubechies wavelets did not give better results
than the old one without any wavelet transformation, moreover, several additional
errors were detected. On the contrary, in case of the rule base made by Haar wavelets,
accurate results were gained for each of the 13 line that produce valid results and
one further line could be assessed, too, as it can be seen in Fig. 6.
In order to evaluate those lines that were previously not to be assessed, this new
(Haar wavelets based) rule base was applied together with the stabilized KH rule in-
terpolation. The 65 test lines were re-processed, thus the predictions became feasible
in the case of all lines. The predictions for the 13 wire pairs which were correctly
Fig. 5 Insertion loss values and the corresponding wavelet transforms. Different performance
classes are indicated by different colors
40 F. Lilik et al.
21,54%
78,46%
70,77%
5 Conclusions
References
1. ITU.: Single-pair high-speed digital subscriber line (SHDSL) transceivers, Technical Recom-
mendation G.991.2, ITU Std., Dec 2003
2. ITU.: Single-pair high-speed digital subscriber line (SHDSL) transceivers, amendment 2, Tech-
nical Recommendation G.991.2 amendment 2, ITU Std., Feb 2005
3. Lilik, F., Nagy, Sz., Kóczy, L.T.: Wavelet Based Fuzzy Rule Bases in Pre-Qualification of
Access Networks Wire Pairs. IEEE Africon 2015, Addis Ababa, Ethiopia, 14–17 Sept 2015
paper P-52, 5 p
4. Balázs, K., Kóczy, L.T.: Constructing dense, sparse and hierarchical fuzzy systems by applying
evolutionary optimization techniques. Appl. Comput. Math. 11(1), 81–101 (2012)
5. Lilik, F., Botzheim, J.: Fuzzy based prequalification methods for EoSHDSL technology. Acta
Tech. Jauriensis Ser. Intelligentia Computatorica 4(1), 135–145 (2011)
6. Lilik, F., Kóczy, L.T.: The determination of the bitrate on twisted pairs by Mamdani infer-
ence method, issues and challenges of intelligent system and computational intelligence. Stud.
Comput. Intell. 530, 59–74 (2014). https://doi.org/10.1007/978-3-319-03206-1_5
7. Daubechies, I.: Ten Lectures on Wavelets. CBMS-NSF Regional Conference Series in Applied
Mathematics 61. SIAM, Philadelphia (1992)
8. Chui, C.K.: An Introduction to Wavelets. Academic Press, San Diego (1992)
9. Christopoulos, Ch., Skodras, A., Ebrahimi, T.: The JPEG2000 still image coding system: an
overview. IEEE Trans. Consum. Electron. 46, 1103–1127 (2000). https://doi.org/10.1109/30.
920468
10. MJ2 format standard, ISO/IEC 15444-3:2002/Amd 2:2003
11. Martucci, S.A., Sodagar, I., Chiang, T., Zhang, Y.-Q.: A zerotree wavelet video coder. IEEE
Trans. Circ. Sys. Video Technol. 7, 109–118 (1997)
12. Chai, B.-B., Vass, J., Zhuang, X.: Significance-linked connected component analysis for
wavelet image coding. IEEE Trans. Image Proc. 8, 774–784 (1999)
13. Montoya Zegarra, J.A., Leiteb, N.J., da Silva, Torres R.: Wavelet-based fingerprint image
retrieval. J. Comput. Appl. Math. 227, 297–307 (2008). https://doi.org/10.1016/j.cam.2008.
03.017
14. Kiely, A., Klimesh, M.: The ICER Progressive Wavelet Image Compressor. IPN Progress
Report 42-155, 15 Nov 2003. http://ipnpr.jpl.nasa.gov/tmo/progressreport/42-155/155J.pdf
15. Grossmann, A., Morlet, J.: Decomposition of hardy functions into square integrable wavelets
of constant shape. SIAM J. Math. Anal. 15, 723–736 (1984)
16. Fourier, J.B.J.: Theorie Analitique de la Chaleur. Didot, Paris (1822)
42 F. Lilik et al.
17. Nagy, Sz., Pipek, J.: On an economic prediction of the finer resolution level wavelet coefficients
in electron structure calculations. Phys. Chem. Chem. Phys. 17, 31558–31565 (2015). https://
doi.org/10.1039/C5CP01214G
18. Kóczy, L.T., Hirota, K.: Approximate reasoning by linear rule interpolation and general ap-
proximation. Int. J. Approximate Reasoning 9, 197–225 (1993). https://doi.org/10.1016/0888-
613X(93)90010-B
19. Kóczy, L.T., Hirota, K.: Interpolative reasoning with insufficient evidence in sparse fuzzy rule
bases. Inf. Sci. 71, 169–201 (1993). https://doi.org/10.1016/0020-0255(93)90070-3
20. Tikk, D., Joó, I., Kóczy, L.T., Várlaki, P., Moser, B., Gedeon, T.D.: Stability of interpolative
fuzzy KH-controllers. Fuzzy Sets Syst. 125, 105–119 (2002). https://doi.org/10.1016/S0165-
0114(00)00104-4
21. Haar, A.: Zur Theorie der Orthogonalen Funktionensysteme. Math. Ann. 69, 331–371 (1910)
Abductive Reasoning on Molecular
Interaction Maps
1 Introduction
This research was partially supported by the French Spanish Laboratory for Advanced Studies
in Information, Representation and Processing (LEA-IREP). Martín Diéguez was supported by
the Centre international de mathématiques et d’informatique (contract ANR-11-LABX-0040-
CIMI).
be modulated by other proteins, which can either enable these reactions or, on the
opposite, inhibit them.
Medical and pharmaceutical researches [15, 19] showed that the break of the
double strand of DNA sometimes appears in a cell as a result of some pathology in
a metabolic pathway, and double strand break (dsb) is a major cause of cancer.
These pathways are used to investigate the molecular determinants of tumor
response in cancers. The molecular parameters include the cell cycle checkpoint,
DNA repair and apoptosis1 pathways [15, 19, 21, 25, 26]. When DNA damage
occurs, cell cycle checkpoints are activated and can rapidly kill the cell by apoptosis
or arrest the cell cycle progression to allow DNA repair before cellular reproduction
or division (see, for instance, the atm-chk2 and atr -chk2 pathways in [26]).
Most of these pathways are formed by an intricate and complex network of chain
reactions, and are often represented in Molecular Interaction Maps (MIM), a human
readable form of the cell cycle checkpoint pathways. MIMs become increasingly
larger and their density is constantly enriched with new information (references, date,
authors, etc.). Although essential for knowledge capitalization and formalization,
MIMs are difficult to use:
– Reading is complex due of the very large number of elements, and reasoning about
the map is even more difficult.
– Using a map to communicate goals is only partially suitable because the represen-
tation formalism requires expertise.
– Maps can contain implicit knowledge, that is taken for granted by one expert, but
is missed by another one.
– Maps can be inconsistent, and these inconsistencies are difficult to detect just by
looking at the map itself.
These problems have been faced from the point of view of nonmonotonic rea-
soning, specially Answer Set Programming (ASP) [13, 14] or action languages [2],
taking advantage of non-monotonictiy and the efficient ASP solvers that are avail-
able. Our approach is based on a reduction to classical logic which allows applying
classical reasoning on such kind of networks.
In this paper we present a method that automatically transforms a MIM into a set
of logical formulas, taking as input the XML files generated by a graphical editor
for biological pathways such as pathvisio [31]. Along this paper we will use, as
examples, subgraphs of the pathway of Fig. 1, which represents the modelling of the
atm-chk2 pathway leading to apoptosis.
The rest of this paper is organized as follows: Sect. 2 introduces the concept of
Molecular Interaction Maps and how they can be translated into a set of logical for-
mulas. Section 3 describes Molecular Interaction Logic, a logic which is capable of
describing and reasoning about general pathways. Section 4 investigates the appli-
cation of deductive as well as abductive reasoning on MIMs. Section 5 presents the
tools used and/or implemented and Sect. 6 investigates the future lines of work.
proteins or conditions. For example, in Fig. 1, atm_ ps1981 blocks the dimerization
of atm into atm_atm, while the double strand break (dsb) of DNA triggers the
production of atm_ ps1981 by atm.
These interactions can be “stacked”: for example, protein p53 can be phosphory-
lated at serine 15 to become p53_ ps15 (see Fig. 1). This reaction is triggered by atm,
but the triggering itself has to be activated by dsb and can be blocked by atm_atm.
Thus, the two main actions (production of a protein or inhibition of a protein) can be
triggered or blocked by a stack of preconditions.
Our first goal is to translate any MIM into a set of logical expressions in order to
perform several automated reasoning tasks such as deduction or abduction. First,
focusing on the diagram of Fig. 2 (which corresponds to a sub-diagram of Fig. 1)
will help getting an intuitive idea of how translation is performed.
Here apoptosis arises when protein p53 is phosphorylated at serine 20 or 15
(instances p53_ ps20 and p53_ ps20 respectively). However, apoptosis would not
happen if the dimer p53_mdm2 is present. Thus the context would be if p53 and
either p53_ ps20 or p53_ ps15 are present and p53_mdm2 is absent then apoptosis
is produced (this example should of course be completed with the rules for producing
the rest of objects in the diagram).
The general form of production relations is displayed in Fig. 3.
Each arrow can be either an activation or an inhibition of the relation it applies to,
and these activations/inhibitions can be stacked on any number of levels. The above
examples give the idea behind the translation: it is a recursive process starting from
the production relation and climbing up the tree. In order to formally describe these
graphs, we define below the concepts of pathway context and pathway formula.
α ::= α P , α Q |P , Q
apoptosis
Abductive Reasoning on Molecular Interaction Maps 47
a1 ,a2 ,...,an b
g1 ,g2 ,...,gn
where P and Q are sets (possibly empty) of propositional variables representing the
conditions of activation () or inhibition () of the reaction. The first part of the
pair is the activation context, the second part is the inhibition context. One, or both
sets can be empty.
For example, the p53 A apoptosis reaction of Fig. 2 would lead to the following
two pathway contexts:
F ::= [α]( p1 ∧ · · · ∧ pn → Pr q) |
[α]( p1 ∧ · · · ∧ pn → In q) |
F∧F
Applied to the example of Fig. 2, the causal pathway formula associated with the
production rule p53 A apoptosis is
48 J.-M. Alliot et al.
In this section the semantics of causal pathway formulas is formally introduced. The
resulting logic, denoted by MIL (Molecular Interaction Logic), extends a previous
work [4, 5] where the MIMs were formalized via first order logic with equality, in
which the pathway contexts were limited to one level of depth. From now on, p
means protein p is present and ¬ p means protein p is absent. Before going into
details, we first provide a method to translate a pathway context into a classical
Boolean expression, which will be used in the definition of the satisfaction relation.
The above expressions define the general forms of A(α) and I (α). If one part of
the context α is empty, then the corresponding part is of course absent in A(α)
and I (α).
The intuitive meaning behind these two evaluations correspond for V1 to the
protein present or absent, and for V2 to the state of the protein resulting from the
chemical reactions in the cell.2
While the first four satisfaction relations are simple to understand, the fifth one
is a little bit trickier according to Definition 3; its meaning is that if the conditions
of activation of α are satisfied, then the reaction represented by F holds. As usual, a
formula F is satisfiable if there is a model (V1 , V2 ) such that (V1 , V2 ) F.
Observation 2 MIL can be characterized by the axioms of classical logic, plus the
axioms:
1. [α]F ↔ (A(α) → F)
2. Pr p → p, if p is produced then p is present
3. In p → ¬ p, if p is inhibited then p is absent
As a result of MIL semantics, the causal pathway formula (3) is logically equivalent
to the conjunction of the following implications:
Observation 3 Any MIM can be transformed into a causal pathway formula, and
every causal pathway formula is equivalent to a boolean composition of:
– propositional variables or their negation
– propositional variables qualified by Pr or In or their negation
2 If the semantics of the modal logic S5 is restricted to have at most two worlds then a strong normal
form in which conjunctions and disjunctions are not in the scope of a modal operator can be found
for this new logic [11]: the pathway causal formulas of MIL verify this condition.
50 J.-M. Alliot et al.
Definition 6 (Completion Axiom from [3]) Let p be an object that can be produced
through different pathways:
C1 → Pr p
···
Cn → Pr p
Every Ci formula represents one pathway leading to p. The completion axiom for
Pr p is the implication Pr p → (C1 ∨ · · · ∨ Cn ).
This axiom means that Pr p is a notion local to the current map and, therefore, p has
to be produced by at least one of the Ci possible pathways. For instance, pathway
rules related to apoptosis in Fig. 2 would lead to the following completion axiom:
For instance, the logical representation of the diagram of Fig. 2 would require the
addition of such axioms for apoptosis, p53_mdm2, p53_ ps20 and p53_ ps15, the
endogenous entities occurring in this map.
3 Ina broadly simplified way, endogenous entities can be considered as “internal variables” or
“output variables” of the model, while exogenous entities are “command variables”.
Abductive Reasoning on Molecular Interaction Maps 51
In this section, deductive and abductive reasoning are used on MIMs in order to find
inconsistencies in the representations and to answer questions about production or
inhibition of proteins in a MIM.
Figure 4 represents the dimerization of atm into atm_atm. When translated into its
logical representation, a SAT-checker finds that this representation is inconsistent.
Why so? When atm and not dsb are present then atm_atm is produced. However,
with dsb, dsb_atm_atm enables the phosphorylation of atm into atm_ ps1981
which then blocks atm_atm, which is inconsistent with the fact that atm_atm is
necessary to produce dsb_atm_atm.
This inconsistency arises because biologists (at least some of them) have some
implicit temporal knowledge about the way these reactions take place, but they do
not represent this knowledge explicitly in MIMs. This shows that if MIMs are to
become a consistent medium for representing proteins interaction, they have to be
enriched with temporal informations.
The production axioms must now be added for all endogenous proteins: p53_ ps15
→ Pr p53_ ps15, p53_ ps20 → Pr p53_ ps20, p53_mdm2 → Pr p53_mdm2, and
apoptosis → Pr apoptosis.
The database is now complete and the system can be queried using abductive
methods (see Sect. 5); for example, the question ?apoptosis gives the answer:
apoptosis ∨ p53_ ps20 ∨ p53_ ps15 ∨ p53. These results are coherent from a bio-
logical point of view. If p_53 is present, then apoptosis occurs, but the same con-
clusion can be obtained from either p53_ ps15 or p53_ ps20, since any of them is
present if p53 is.
The diagram of Fig. 2 is now modified by adding a new protein, chk2, which is
required to activate the production of p53_ ps20.
This new representation (Fig. 5) implies that rule (9) becomes p53 ∧ chk2 →
Pr p53_ ps20 and the corresponding completion axiom (13) becomes: Pr p53_ ps20
→ p53 ∧ chk2.
When applying abductive reasoning on the new diagram, the answer is pretty
unexpected:
p53
apoptosis
Now p53 alone is not enough for apoptosis anymore; ( p53 and chk2), or ( p53 and
not mdm2) are needed for apoptosis. It is not immediately clear why the simple
introduction of chk2 in the pathway labelled as pathway 2 changes the answer. This
comes from the fact that, in this diagram, p53_ ps15 does not block the production
of p53_mdm2 explicitly. Therefore, if both p53 and mdm2 are present but chk2 is
not, then p53_mdm2 is produced in pathway 2 blocking apoptosis in pathway 1.
As a last, complicated test, the complete diagram in Fig. 1 which represents the whole
atm − chk2 pathway was used. The logical representation of this diagram requires
around 80 clauses. The abductive query (?apoptosis) was performed to find all
conditions on exogenous variables that lead to apoptosis. The answer (( p53 ∧ dsb ∧
atm) ∨ ( pml ∧ chk2 ∧ dsb ∧ atm)) are both correct and a little bit surprising. They
clearly correspond to the paths leading to apoptosis1 and apoptosis3. The reason
why the conditions leading to apoptosis2 do not appear becomes clear when solving
?apoptosis2 alone. The answer to this question (( p53 ∧ chk2 ∧ dsb ∧ atm)) is
subsumed by the one leading to apoptosis1.
Abductive queries give also the states of all the endogenous variables in the graphs
which lead to apoptosis. They can be an invaluable tool for biologists on complex
MIMs, where establishing such results by hand is both tedious and error-prone.
5 Implementation
All the examples presented here were written using Pathvisio, a public-domain MIM
editing software.
A parser was developed to automatically translate the MIM from the XML gener-
ated by Pathvisio, add the formulas generated by the completion axioms and trans-
form the formulas into Conjunctive Normal Form.
As demonstrated above, using Observation 4 we transform each set of clauses of
our language in a set of classical clauses, allowing the use of both deduction and
abduction algorithms for classical logic. Regarding deduction and SAT-checking,
several efficient tools are available, such as the minisat solver [9].
54 J.-M. Alliot et al.
References
1. Alberti, M., Gavanelli, M., Lamma, E., Mello, P., Torroni, P.: The SCIFF abductive proof-
procedure. In: AI*IA’05, p. 135147 (2005)
2. Baral, C., Chancellor, K., Tran, N., Tran, N., Joy, A., Berens, M.: A knowledge based approach
for representing and reasoning about signaling networks. Bioinformatics 20 (suppl 1), i15–i22
(2004)
Abductive Reasoning on Molecular Interaction Maps 55
3. Clark, K.L.: Negation as failure. In: Logic and Databases, pp. 293–322. Plenum Press (1978)
4. Demolombe, R., Fariñas del Cerro, L., Obeid, N.: Logical model for molecular interactions
maps. In: Logical Modeling of Biological Systems, pp. 93–123. Wiley, New York (2014)
5. Demolombe, R., Fariñas del Cerro, L., Obeid, N.: Translation of first order formulas into ground
formulas via a completion theory. J. Appl. Logic (), to appear
6. Demolombe, R., Fariñas del Cerro, L.: An inference rule for hypothesis generation. In: IJCAI’91
(1991)
7. Demolombe, R., Fariñas del Cerro, L.: Information about a given entity: from semantics towards
automated deduction. J. Logic Comput. 20(6), 1231–1250 (2010)
8. Denecker, M., Kakas, A.: Abduction in logic programming. In: Computational Logic: Logic
Programming and Beyond, pp. 402–436 (2002)
9. Een, N., Srensson, N.: An extensible sat-solver. In: SAT’03, pp. 502–518 (2003)
10. Erwig, M., Walkingshaw, E.: Causal reasoning with neuron diagrams. In: VLHCC ’10, pp.
101–108 (2010)
11. Fariñas del Cerro, L., Herzig, A.: Contingency-based equilibrium logic. In: LPNMR’11, pp.
223–228 (2011)
12. Fariñas del Cerro, L., Inoue, K. (eds.): Logical Modeling of Biological Systems. Wiley, New
York (2014)
13. Gebser, M., Guziolowski, C., Ivanchev, M., Schaub, T., Siegel, A., Thiele, S., Veber, P.: Repair
and prediction (under inconsistency) in large biological networks with answer set programming.
In: KR’10 (2010)
14. Gebser, M., Schaub, T., Thiele, S., Veber, P.: Detecting inconsistencies in large biological net-
works with answer set programming. Theory Pract. Logic Program. 11(2–3), 323–360 (2011)
15. Glorian, V., Maillot, G., Poles, S., Iacovoni, J.S., Favre, G., Vagner, S.: Hur-dependent loading
of mirna risc to the mrna encoding the ras-related small gtpase rhob controls its translation
during uv-induced apoptosis. Cell Death Differ. 18(11), 1692–1701 (2011)
16. Inoue, K.: Linear resolution for consequence finding. Artif. Intell. 56(2–3), 301–353 (1992)
17. Jackson, P.: Computing prime implicates incrementally. In: CADE’92, pp. 253–267 (1992)
18. Kean, A., Tsiknis, G.: An incremental method for generating prime implicants/implicates. J.
Symbolic Comput. 9, 185–206 (1990)
19. Kohn, K.W., Pommier, Y.: Molecular interaction map of the p53 and mdm2 logic elements,
which control the off-on swith of p53 response to dna damage. Biochem. Biophys. Res. Com-
mun. 331(3), 816–27 (2005)
20. Kohn, K.W., Aladjem, M.I., Weinstein, J.N., Pommier, Y.: Molecular interaction maps of
bioregulatory networks: a general rubric for systems biology. Mol. Biol. Cell. 17(1), 1–13
(2006)
21. Lee, W., Kim, D., Lee, M., Choi, K.: Identification of proteins interacting with the catalytic
subunit of pp2a by proteomics. Proteomics 7(2), 206–214 (2007)
22. Muggleton, S., Bryant, C.H.: Theory completion using inverse entailment. In: ILP’00, pp.
130–146 (2000)
23. Nabeshima, H., Iwanuma, K., Inoue, K., Ray, O.: SOLAR: an automated deduction system for
consequence finding. AI Commun. 23(2–3), 183–203 (2010)
24. Nuffelen, B.V.: A-system: problem solving through abduction. BNAIC01 Sponsors 1, 591–596
(2001)
25. Pei, H., Zhang, L., Luo, K., Qin, Y., Chesi, M., Fei, F., Bergsagel, P.L., Wang, L., You, Z., Lou,
Z.: MMSET regulates histone H4K20 methylation and 53BP1 accumulation at DNA damage
sites. Nature 470(7332), 124–128 (2011)
26. Pommier, Y., Sordet, O., Rao, V.A., Zhang, H., Kohn, K.W.: Targeting chk2 kinase: molecular
interaction maps and therapeutic rationale. Curr. Pharm. Des. 11(22), 2855–72 (2005)
27. Ray, O., Kakas, A.: ProLogICA: a practical system for abductive logic programming. In:
Proceedings of the 11th International Workshop on Non-monotonic Reasoning, pp. 304–312
(2006)
28. Ray, O., Whelan, K., King, R.: Logic-based steady-state analysis and revision of metabolic
networks with inhibition. In: CISIS’10, pp. 661–666 (2010)
56 J.-M. Alliot et al.
29. Reiser, P.G., King, R.D., Kell, D.B., Muggleton, S., Bryant, C.H., Oliver, S.G.: Developing a
logical model of yeast metabolism. Electron. Trans. Artif. Intell. 5, 233–244 (2001)
30. Rougny, A., Froidevaux, C., Yamamoto, Y., Inoue, K.: Analyzing SBGN-AF Networks Using
Normal Logic Programs. In: Logical Modeling of Biological Systems, pp. 44–55. Wiley, New
York (2014)
31. van Iersel, M.P., Kelder, T., Pico, A.R., Hanspers, K., Coort, S., Conklin, B.R., Evelo, C.:
Presenting and exploring biological pathways with pathvisio. BMC Bioinform., p. 399 (2008)
Efficient Unfolding of Fuzzy Connectives
for Multi-adjoint Logic Programs
Abstract During the last decade we have designed several tools for assisting the
development of flexible software applications coded with a promising language in
the fuzzy logic programming area. In the so-called multi-adjoint logic programming
approach, a set of logic rules are assembled with a set of fuzzy connective definitions
(whose truth functions are defined as functional rules) for manipulating truth degrees
beyond the simpler case of {true,false}. Moreover, we have recently provided opti-
mization techniques by reusing some variants of program transformation techniques
based on unfolding which have been largely exploited in the pure functional -not
fuzzy- setting for enhancing the behavior of such operators. In this paper we exper-
imentally show the benefits of using the new c-unfolding transformation applied on
fuzzy connectives and how to improve the efficiency of the proper unfolding pro-
cess by reusing the very well-known concept of dependency graph. Moreover, we
accompany our technique with cost analysis and discussions on practical aspects.
1 Introduction
Although logic programming [24] has been widely used as a formal method for prob-
lem solving and knowledge representation, fuzzy logic programming has emerged as a
growing research area for incorporating techniques or constructs based on fuzzy logic
This work has been partially supported by the EU (FEDER), the State Research Agency (AEI)
and the Spanish Ministerio de Economía y Competitividad under grant TIN2016-76843-C4-2-R
(AEI/FEDER, UE).
to explicitly deal with uncertainty and approximated reasoning. Most fuzzy logic
languages developed during the last decades implement (extended versions of) the
resolution principle introduced by Lee [21], such as Elf-Prolog [14], F-Prolog [23],
generalized annotated logic programming [19], Fril [7], MALP [27], FASILL [15],
the QLP scheme of [37] and the many-valued logic programming language of [38].
In this paper we focus on the so-called multi-adjoint logic programming approach
MALP [25–27], a powerful and promising proposal in the area of fuzzy logic pro-
gramming for which we have developed the FLOPER system (see [31, 36] and visit
the Web site http://dectau.uclm.es/floper/) as well as several techniques for develop-
ing, optimizing and tuning fuzzy applications [10, 17, 33]. Intuitively speaking, logic
programming is extended with a multi-adjoint lattice L of truth values (typically, a
real number between 0 and 1), equipped with a collection of adjoint pairs &i , ←i
and connectives: implications, conjunctions, disjunctions, and other operators called
aggregators, which are interpreted on this lattice.
In Sect. 2 we explain both the syntax and operational semantics of the MALP
language where, in essence, to solve a MALP goal, i.e., a query to the system plus
a substitution (initially the empty substitution, denoted by id), a generalization of
the classical modus ponens inference rule called admissible steps are systematically
applied on atoms in a similar way to classical resolution steps in pure logic program-
ming, thus returning a state composed by a computed substitution together with an
expression where all atoms have been exploited. Next, this expression is interpreted
under a given lattice, hence returning a pair truth degree; substitution which is
the fuzzy counterpart of the classical notion of computed answer used in pure logic
programming.
In Sect. 3 we collect from [29] our technique for reducing the complexity of con-
nectives (also alleviating the computational cost of derivations) by safely removing all
the intermediate calls performed on the equations defining the behavior of such con-
nectives. We show that this process can be easily described in terms of “unfolding”,
a well-known, widely used, semantics-preserving program transformation operation
which in most declarative paradigms is usually based on the application of computa-
tion steps on the body of program rules (in [10, 16, 17] we describe our experiences
regarding the unfolding of fuzzy logic programs). The novelty of our approach is that
it is the first time that unfolding is not applied to program rules, but to connective
definitions, maintaining the same final goal, i.e., generating more efficient code. The
main three goals addressed in this paper are:
1. The advantages of the unfolding transformation adapted to connective definitions
in our fuzzy setting are experimentally checked at the end of Sect. 3.
2. Moreover, we reuse in Sect. 4 some techniques based on dependency graphs for
improving the proper transformation process as much as possible, also including
computational cost analysis for our resulting algorithm.
3. In Sect. 5 we discuss some practical issues by connecting our techniques with
several tools implemented in our research group.
Finally, before concluding in Sect. 7, a few hints on related work—which also help
to motivate our approach—are presented in Sect. 6.
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 59
This section summarizes the main features of multi-adjoint logic programming (see
[25–27] for a complete formulation of this framework). We work with a first order
language, L, containing variables, constants, function symbols, predicate symbols,
and several (arbitrary) connectives to increase language expressiveness: implication
connectives (←1 , ←2 , . . .); conjunctive operators (denoted by &1 , &2 , . . .), disjunc-
tive operators (|1 , |2 , . . .), and hybrid operators (usually denoted by @1 , @2 , . . .),
all of them are grouped under the name of “aggregators” or directly “connectives”.
Aggregation operators are useful to describe/specify user preferences. An aggrega-
tion operator, when interpreted as a truth function, may be an arithmetic mean, a
weighted sum or in general any monotone application whose arguments are values
of a complete bounded lattice L. For example, if an aggregator @ is interpreted
as [[@]](x, y, z) = (3x + 2y + z)/6, we are giving the highest preference to the
first argument, then to the second, being the third argument the least significant.
Although these connectives are binary operators, we usually generalize them as
functions with an arbitrary number of arguments. So, we often write @(x 1 , . . . , xn )
instead of @(x1 , . . . , @(xn−1 , xn ), . . .). By definition, the truth function for an n-
ary aggregation operator [[@]] : L n → L is required to be monotonous and fulfills
[[@]](, . . . , ) = , [[@]](⊥, . . . , ⊥) = ⊥.
Additionally, our language L contains the values of a multi-adjoint lattice,
L , , ←1 , &1 , . . . , ←n , &n , equipped with a collection of adjoint pairs ←i , &i ,
where each &i is a conjunctor which is intended to the evaluation of modus ponens
[27]. In general, L may be the carrier of any complete bounded lattice but, for read-
ability reasons, in the examples we shall select L as the set of real numbers in the
interval [0, 1]. A L-expression is a well-formed expression composed by values and
connectives of L, as well as variable symbols and primitive operators (i.e., arithmetic
symbols such as ∗, +, min, etc.). In what follows, we assume that the truth func-
tion of any connective @ in L is given by its corresponding connective definition,
that is, an equation or rewriting rule of the form @(x 1 , . . . , xn ) = E, where E is a
L-expression not containing variable symbols apart from x1 , . . . , xn .
A rule is a formula H ←i B, where H is an atomic formula or atom (usually
called the head) and B (which is called the body) is a formula built from atomic
formulas B1 , . . . , Bn —n ≥ 0—, truth values of L, conjunctions, disjunctions and
aggregations. A goal is a body submitted as a query to the system. Roughly speaking,
a multi-adjoint logic program is a set of pairs R; α (we often write R with α),
where R is a rule and α is a truth degree (a value of L) expressing the confidence of
a programmer in the truth of the rule R. By abuse of language, we sometimes refer
a tuple R; α as a “rule”.
The procedural semantics of the multi-adjoint logic language L can be thought as
an operational phase (based on admissible steps) followed by an interpretive one. In
the following, C[A] denotes a formula where A is a sub-expression which occurs in
the—possibly empty—context C[]. Moreover, C[ A/A ] means the replacement of A
by A in context C[], whereas Var (s) refers to the set of distinct variables occurring
60 P. J. Morcillo and G. Moreno
in the syntactic object s, and θ [Var (s)] denotes the substitution obtained from θ by
restricting its domain to Var (s).
where the labels luka, godel and prod mean respectively for Łukasiewicz logic,
Gödel logic and product logic, that is,
Now, we can generate the following admissible derivation (we underline the selected
atoms in each step):
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 61
AS1R1
p(X ); id
AS2R2
&godel (0.9, &prod (|luka (q(X 1 ), 0.6), r (X 1 ))); {X/ X 1 }
AS2R3
&godel (0.9, &prod (|luka (0.8, 0.6), r (a))); {X/a, X 1 /a}
&godel (0.9, &prod (|luka (0.8, 0.6), 0.7)); {X/a, X 1 /a, X 2 /a}
Here, the admissible computed answer (a.c.a.) is the pair: &godel (0.9,&prod
(|luka (0.8, 0.6), 0.7)); {X/a}.
If we exploit all atoms of a goal, by applying admissible steps as much as needed dur-
ing the operational phase, then it becomes a formula with no atoms (a L-expression)
which can be then directly interpreted w.r.t. lattice L. Although in [27] this last pro-
cess is implicitly included in a definition similar to the previous one for describing
the intended notion of fuzzy computed answer, here we prefer to model it as a new
computational process (transition system) by applying the following definition we
initially presented in [18] (as we will see in further sections, the cost measures pro-
posed in this paper and previous ones [30, 32], are strongly related with the behaviour
and detailed definition of interpretive step):
Definition 4 Let P be a program and Q; σ an a.c.a., that is, Q is a goal not
containing atoms (i.e., a L-expression). An interpretive derivation is a sequence
IS
Q; σ ∗ Q ; σ . When Q = r ∈ L, being L , the lattice associated to P, the
state r ; σ is called a fuzzy computed answer (f.c.a.) for that derivation.
IS
&godel (0.9, &prod (|luka (0.8, 0.6), 0.7)); {X/a}
IS
&godel (0.9, &prod (1, 0.7)); {X/a}
IS
&godel (0.9, 0, 7); {X/a}
0.7; {X/a}.
As we said in the previous section, connective definitions are equations (or rewrit-
ing rules) of the form @(x1 , . . . , xn ) = E, where E is a L-expression which might
contain variable symbols in the set {x1 , . . . , xn }, as well as values, primitive opera-
tors and connectives of a multi-adjoint lattice L. The use of connectives inside the
definition of other connectives is a powerful expressive resource useful not only
for programmers interested in describing complex aggregators, but it also plays an
important role in fuzzy transformation techniques such as the fold/unfold framework
we have described in [10, 16–18]. Consider for instance, the following connective
definition: @complex (x1 , x2 ) = &prod (|luka (x1 , 0.6), x2 ). This hybrid aggregator
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 63
was used in [30, 32] (with slight modifications) for pointing out some observed
discrepancies when measuring the interpretive cost associated to the execution of
MALP programs.
AS1R4
p(X ); id
AS2R2
&godel (0.9, @complex (q(X 1 ), r (X 1 )); {X/ X 1 }
AS2R3
&godel (0.9, @complex (0.8, r (a))); {X/a, X 1 /a}
IS
&godel (0.9, @complex (0.8, 0.7)); {X/a, X 1 /a, X 2 /a}
IS
&godel (0.9, 0.7); {X/a, X 1 /a, X 2 /a}
0.7; {X/a, X 1 /a, X 2 /a}
Note that, since we have exploited the same atoms with the same rules (except for
the first steps performed with the equivalent rules R1 and R4 , respectively) in both
derivations, then O c (D1 ) = O c (D1∗ ) = 3. However, although connectives &prod and
|luka have been evaluated in both derivations, in D1∗ such evaluations have not been
explicitly counted as interpretive steps, and consequently they have not been added
to increase the interpretive cost measure I c . This unrealistic situation is reflected by
the abnormal result: I c (D1 ) = 3 > 2 = I c (D1∗ ). In [30, 32] we have described two
different techniques (based respectively on a redefinition of the notion of interpretive
step and in the introduction of the concept of “weight of a connective”) evidencing
64 P. J. Morcillo and G. Moreno
that the interpretive cost of derivation D1∗ is not only lower, but even greater than
derivation D1 . The main reason is that complex connective definitions involving calls
to other aggregators consume more computational resources than other connectives
which only evaluate primitive operators.
The previous example motivates the following definition, which in essence describes
a technique based on classical unfolding transformations for simplifying, when pos-
sible, connective definitions by “unnesting” unnecessary calls to other connectives.
We assume here that the rules (equations) describing connective definitions are taken
renamed apart (at least one of them) before applying an unfolding step, as it is also
usual with program rules in many declarative transformation tasks.
... and remembering that |luka (x1 , x2 ) = min(1, x1 + x2 ), then, we can unfold con-
nective @complex w.r.t. connective |luka as follows:
– Firstly, we generate the “matcher” between the call |luka (x1 , 0.6) appearing in the
“rhs” (right hand side) of the first equation rule and the “lhs” (left hand side) of
the second rule |luka (x1 , x2 ), thus producing links x1 /x1 and x2 /0.6.
– Next, we apply both bindings to the rhs of the second rule, obtaining the L-
expression min(1, x1 + 0.6).
– Then, this L-expression is used to replace the original call to |luka in the rhs of
the first rule, producing &prod (min(1, x1 + 0.6), x2 ).
– Finally, this last L-expression conforms the rhs of the new connective definition
for @complex , that is: @complex (x1 , x2 ) = &prod (min(1, x1 + 0.6), x2 ).
Following the same method, but performing now the c-unfolding of @complex w.r.t.
&prod whose connective definition is &prod (x1 , x2 ) = x1 ∗ x2 , we obtain the final
rule defining @complex with the following shape @complex (x1 , x2 ) = min(1, x1 +
0.6) ∗ x2 . Note that the new connective definition is just a simple arithmetic expres-
sions involving primitive operators but no calls to other connectives, as wanted. From
now on, this improved definition will be referred as @unfolded .
To finish this section, in Table 1 we show the benefits of using c-unfolding by means
of a experimental evaluation performed on a desktop computer equipped with an
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 65
Table 1 Evaluating original and unfolded fuzzy connectives: runtimes and speed-up
MN 100 1000 10,000
100 0.04 / 0.03 = 1.33 0.31 / 0.3 = 1.03 2.91 / 2.89 = 1.01
1000 0.11 / 0.03 = 3.67 0.38 / 0.3 = 1.27 3.03 / 2.93 = 1.03
10000 0.76 / 0.04 = 19 1.02 / 0.3 = 3.4 3.62 / 2.92 = 1.24
i3-2310M CPU @ 2.10 GHz and 4,00 GB RAM. We consider an initial com-
plex connective whose definition contains (in a direct or indirect way) N calls to
other connectives and requires the evaluation of M primitive operators like min,
∗, +, and so on. More exactly, assuming that a connective @ directly evaluates m
primitive operators and performs n direct calls to connectives @1 , . . . , @n , then,
we can compute values M and N by means of the auxiliary functions oper s and
calls as follows: M = oper s(@) = m + oper s(@1 ) + · · · + oper s(@n ) and N =
calls(@) = n + calls(@1 ) + · · · + calls(@n ). For instance, if we try to compute
the values M and N for connective @complex (x1 , x2 ) = &prod (|luka (x1 , 0.6), x2 )
in Example 4, we have that m = 0, n = 2 and since |luka (x1 , x2 ) = min(1, x1 + x2 )
and &prod (x1 , x2 ) = x1 ∗ x2 , then oper s(|luka ) = 2 and oper s(&prod ) = 1, while
calls(|luka ) = calls(&prod ) = 0, which implies that M = oper s(@complex ) =
m + oper s(|luka ) + oper s(&prod ) = 0 + 2 + 1 = 3 and N = calls(@complex ) =
n + calls(|luka ) + calls(&prod ) = 2 + 0 + 0 = 2. Remember that after applying
c-unfolding, the improved connective definition does not perform calls to any
other connective, but evaluates the same number of primitive operators M (in fact,
observe that in the improved definition @complex (x1 , x2 ) = (min(1, x1 + 0.6)) ∗ x2
of Example 4 we have that M = 3 and N = 0, as wanted).
Each row in Table 1 refers to a different value for N while each column indicates
an alternative assignment to M. Both parameters vary according to values 100, 1000
and 10,000. Each case has been executed 1000 times and the content of each cell has
the form “runtime-original-connective / runtime-unfolded-connective = speed-up”,
where execution times are expressed in milliseconds. Note that the speed up in the
cells of the first row are not significant due to the fact that the number of connective
calls N is never greater than the number of primitive operators M. On the contrary,
in the last row, since N is always greater or equal than M, we obtain good ranges
of speed up. In particular, this measure is 19 in the leftmost cell due to the fact that
c-unfolding removes all the connective calls which caused the low efficiency of the
initial connective definition.
The use of “graphs” (and many different extensions/variants of this formal concept)
as an auxiliary data structure helping to analyze the behaviour of systems/programs
66 P. J. Morcillo and G. Moreno
at several levels, also taking profit in practice of its deep mathematical background.
For instance, and simply focusing on termination topics in declarative programming
(which has somehow influenced our recent research interest), the notions of depen-
dency graphs and size-change graphs have been well reported in [5, 22]. In this
section, we will use the first concept for efficiently guiding the c-unfolding process.
Our experiences in fuzzy fold/unfold transformations [10, 16, 17], reveal us that
drastic situations associated to degenerated transformation sequences might even-
tually produce highly nested definitions of connectives. For instance, assume the
following sequence of (extremely inefficient) connective definitions:
When trying to solve two expressions of the form @100 (0.9, 0.8) and @0 (0.9, 0.8),
we obtain the same result 0.72, but the effort needed to solve the first expression is
very high (due to the 100 avoidable calls to auxiliary connectives) compared with
the second expression (which simply evaluates the arithmetic operator ∗).
Fortunately, by systematically performing c-unfolding on the previous connec-
tives, this problem is successfully solved in a simple way: after applying a c-unfolding
step on aggregator @100 we obtain @100 (x1 , x2 ) = @98 (x1 , x2 ), which admits a new
c-unfolding process to become @100 (x1 , x2 ) = @97 (x1 , x2 ), and following this trail,
after applying the final one-hundredth c-unfolding step, we reach the desired connec-
tive definition @100 (x1 , x2 ) = x1 ∗ x2 . Of course, the transformation process does
not finish here, because we also need to rearrange the shape of all the remaining
connective definitions. So, for each aggregator @i , 0 ≤ i ≤ 100, we need exactly i
c-unfolding steps to achieve the appropriate connective definition.
However, there exist a second, much more intelligent alternative to highly reduce
the number of transformation steps needed to obtain the same final set of improved
connective definitions. In our example, the idea is to proceed just in the inverse
order than previously. So, since @0 does not admit unfolding, we proceed with
@1 , whose connective definition becomes @1 (x1 , x2 ) = x1 ∗ x2 after just a single
c-unfolding step. Now, we take profit of this improved definition when unfolding
@2 , since in just a unique (not two) c-unfolding step we obtain the optimal definition
@2 (x1 , x2 ) = x1 ∗ x2 . Note that the benefits of this last process, are also inherited
when transforming @3 , @4 and so on. So, the advantages obtained after applying each
c-unfolding on a different connective, are “propagated” to the remaining connectives
being improved, which implies that we simply need one hundred transformation steps
to optimize the definitions of the whole set of connectives.
In order to identify in a systematic way the best ordering for performing c-
unfolding operations on connectives, we firstly construct the dependency graph of
a multi-adjoint lattice L associated to a given program P, i.e., a directed graph that
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 67
contains the connective symbols as nodes and an edge from connective @ to aggre-
gator @ for each connective definition in L of the form @(x1 , . . . , xn ) = E, where
the L-expression E contains a call to @ . Given an edge from node @ to node @ , we
denote it as an out-edge of @ and as an in-edge of @ . For instance, the dependency
graphs associated to all the connectives seen so far are:
∨L ←− @∗ −→ &P
@100 −→ @99 −→ . . . −→ @1 −→ @0
As we are going to see, the use of dependency graphs will largely help us to decide
when to unfold each connective in order to minimize the number of transformation
steps. Anyway, before doing this, it is important to note that the construction of such
graphs constitute a fast way to detect possibly abnormal connective definitions, that
is, those ones involved on cycles in the graph (because their further evaluation might
fall in infinite loop). Fortunately, the presence of cycles is not usual in the dependency
graphs associated to connective definitions.
As formalized in the algorithm of Fig. 2, when selecting a connective to apply
c-unfolding, we give priority to those ones without out-edges (and obviously not
belonging to cycles), as occurs in our examples with nodes labeled with ∨L , &P and
@0 , which in our particular case do not need c-unfolding because their definitions do
not perform calls to other aggregators. Once a concrete connective has been selected
and then unfolded as much as possible (and hence, its definition has been completely
improved by removing all its auxiliary calls), then the proper node as well as all
its in-edges (remember that it has not associated out-edges) are removed from the
graph. The process is iterated as much as needed until the dependency graph becomes
empty. For instance in our example, once removed nodes ∨L , &P and @0 , the new
candidates are nodes @∗ and @1 . The first one is unfolded w.r.t. ∨L and &P and
then removed, whereas the second one is dropped out after being unfolded w.r.t.
@0 . Then the process continues with @2 , next @3 and so on, being @100 the last
In this section we establish the feedback and synergies among several independent
tools developed in our research group during the last decade. Whereas in [9] we
present a recent graphical tool for assisting the design of lattices of truth degrees (see
Fig. 3), the functional-logic transformation system SYN T H enables the unfolding-
based optimization of their connective definitions in order to improve the compu-
tational behaviour of those fuzzy logic programs developed with the FLOPER
environment.
The transformation system SYN T H [4, 35] was initially conceived for optimiz-
ing functional logic programs (Curry [11]), and then also used to manipulate pure
functional programs (Haskell [13]). The tool implements five basic transformation
rules (unfolding, folding, abstraction, definition introduction and definition elimina-
tion) and two automatic transformation strategies (composition and tupling), in order
to generate efficient sets of rewriting rules coded with Haskell/Curry syntax.1
For the purposes of the present work, we simply need to consider the unfolding oper-
ation of the SYN T H system for being applied on rewriting rules modeling fuzzy
connectives, as occurs with the following ones of our running example:
Here, the connectives |luka , &prod and @complex are respectively denoted by
or_luka, and_prod and agr_complex; we use the names min, add and prod
for referring to the primitive arithmetic operators min, + and ∗, respectively; and
finally A and B are variable symbols. Once the previous set of connective defini-
tions is loaded into the SYN T H system, it conforms the initial program in the
transformation sequence denoted by “asc 0”. Next, we must select the third rule
and click on the unfolding button, thus obtaining the next program “asc 1”
1 In essence, both languages share the same syntax, but they have a different computational behaviour
since Curry extends with extra logic features the pure functional dimension of Haskell.
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 69
where the selected rule is replaced by the new fourth one agr_complex A B =
and_prod (min 1 (add A (0,6))) B. Finally, this last rule admits a final
unfolding step to reach our intended final definition for agr_complex (where no
calls to other connectives appear), represented by the fifth rule agr_complex A B
= prod (min 1 (add A (0,6))) B in program “asc 2”. Figure 4 shows
the original and final program of the transformation sequence, where the initial and
improved definitions of agr_complex appear darked in blue.
On the other hand, our “Fuzzy LOgic Programming Environment for Research”
FLOPER, is able to trace the execution of goals with respect to a given MALP pro-
gram and its associated lattice, by drawing derivation trees as the ones shown in
Figs. 1 and 5. When we choose option “ismode = small” then the system is
able to detail the set of connective calls performed along a derivation (sis1 steps)
as well as the set of primitive operators (sis2 steps) evaluated in the same deriva-
tion. Before explaining the three derivations collected on the tree drawn in Fig. 5,
remember that our MALP program looks like:
where @unfolded represents the improved definition (where no calls to other con-
nectives are performed) of @complex .
In the figure, each derivation starts with an admissible step exploiting a different
MALP rule defining predicate p (i.e., R1 , R4 and R5 , respectively) and continues
with rules R2 and R3 in all cases until finishing the admissible phase. From here, the
interpretive phase differs in each derivation since, even when all them evaluate exactly
the same set of primitive operators (so, the number of sis2 steps do coincide), they
perform a different number of direct/indirect calls to connectives (represented by
sis1 steps). So, the left-most branch in Fig. 5 is shorter than the branch in the
middle of the tree (as we have largely explained in the previous section), but note
that the right-most branch, which makes use of the connective whose definition has
been improved by means of our c-unfolding technique, is just the shortest one in the
whole figure, as wanted.
We have just seen by means of a very simple example that the “quality” of the con-
nective definitions accompanying a MALP program directly reverts on the efficiency
of the applications coded with this fuzzy language. Thus, the unfolding technique
applicable on fuzzy connectives described in this paper, is intended to play an impor-
tant role in the performance of large scale programs developed with our FLOPER
environment, as it is the case of the real-world application we have recently developed
with FLOPER in the field of the semantic web [2, 3] (see Fig. 6).
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 71
6 Related Work
The ideas managed in the previous sections have been mainly inspired by our previous
studies and experiences in the following two topics: fold/unfold transformations in
declarative programming and cost measures in multi-adjoint logic programming ([28,
30, 32]).
Focusing on primitive functional programs, the pioneer work [8] initiates a fer-
tile tradition in program optimization techniques based on fold/unfold transforma-
tions, which has highly attracted a wide audience in the declarative programming
research community during the last three decades (see the first introduction into logic
programming in [39], and then our adaptations to functional logic programming
72 P. J. Morcillo and G. Moreno
in [4, 35] and fuzzy logic programming in [10, 16–18]). This approach is based
on the construction, by means of a strategy (heuristic), of a sequence of equivalent
programs—called transformation sequence and usually denoted by P 0 , . . . , P n such
that P n is more efficient than P 0 —where each program P i is obtained from the pre-
ceding ones P 0 , . . . , P i−1 by using an elementary transformation rule. The essential
rules are folding and unfolding, i.e., contraction and expansion of sub-expressions
of a program using the definitions of this program (or a preceding one).
Example 5 Consider for instance the following set of rewriting rules describing
classical operations for concatenating two and three lists via the app and double
function symbols, respectively (here we use the constructor symbols “[ ]” and “:” to
model empty and non-empty lists, respectively):
R1 : app([ ], L) =L
R2 : app(H : T, L) = H : app(T, L)
R3 : double(A, B, C) = app(app(A, B), C)
Now, observe that the underlined expression in R6 coincides with the body of the
right hand side of rule R3 , and thus, such expression admits a “contraction” in
terms of “double”. This is just what the final folding step produces, thus generating
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 73
R1 : app([ ], L) =L
R2 : app(H : T, L) = H : app(T, L)
R4 : double([ ], B, C) = app(B, C)
R7 : double(H : A, B, C) = H : double(A, B, C)
On the other hand, in [10] we have developed a much more involved fuzzy fold/unfold
transformation system especially tailored for the MALP framework, whose advan-
tages were illustrated with the following program:
where the new aggregators used in the body of rules R16 and R18 are
defined as @1 (x1 , x2 ) = &prod (0.8, x1 ), &prod (0.95, x2 ) and @2 (x1 , x2 ) =
&godel (x1 , x2 ). In this fuzzy setting, apart from the folding/unfolding operations,
it is mandatory to introduce two new transformations called “tupled definition intro-
duction” and “aggregation”, being this last operation strongly connected with the
notions introduced in this paper, as we are going to see.
As in our previous example, Burstall and Darlington considered in [8] a set of equa-
tions defining functions (also using variables and constructor symbols to build data
structures) as functional programs, with a shape very close to the connective defini-
tions used in this paper, by simply associating their notions of function symbols and
constructor symbols (or more properly, “defined function symbols” and “constructor
function symbols”, respectively), to our concepts of connective symbols and primi-
tive operators, respectively. In fact, our equations describing connectives can be seen
as a particular case of their equations defining program rules, since in the left hand
sides of connective definitions we only allow variables as arguments (they can also
use terms built with constructors).
74 P. J. Morcillo and G. Moreno
For this reason, our notion of c-unfolding does not only accomplish with the
original unfolding definition of [8] (and further, more complex extensions coping
with modern functional and functional-logic programs expressed as Term Rewriting
Systems, TRS’s [6, 12, 20]), but it is even easier, since we simply need a restricted
kind of equations to model connective definitions. Hence, apart from our initial, novel
use of unfolding described in this paper, we think that in the near future it would be
possible to extrapolate to our fuzzy setting many results on transformation techniques
(including folding) with a functional taste, but focusing now on the optimization of
connective definitions.
At this point, we think that it is mandatory to explain why in our initial fuzzy
fold/unfold transformation system described in [10, 18, 34] there exists the risk of
generating artificial and inefficient connective definitions when optimizing multi-
adjoint logic programs. The problem emerges in the fuzzy unfolding transformation
since this operation introduces a great amount of connectives and truth degrees on
unfolded program rules which obstruct the application of further folding steps (as
we have done in Example 5 when generating efficient recursive definitions).
Example 6 Such effects are easy to see even in the propositional case. So, in order
to unfold a fuzzy program rule like:
where the underlined elements confirm our previous comment, that is, the introduc-
tion of extra truth degrees and connectives on the body of unfolded rules.
It is important to note that the existence of such “noisy” elements only emerge in the
fuzzy setting (in contrast with other “crisp” languages) and, for manipulating them,
we require the application of auxiliary techniques which, as we are going to see, will
produce artificial connective definitions: the main motivation of this paper is just
the optimization of such connective definitions by following standard declarative
techniques classically proposed for the optimization of program rules.
So, if we revise the so-called “aggregation transformation rule” described in [10],
we observe that its main goal is to simplify the shape of program rules, by moving
several connective calls from their bodies to the definition of new connective symbols,
in order to give chances for further folding steps to proceed.
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 75
Example 7 Having a look to previous examples, the effects produced by the “aggre-
gation transformation rule” are:
– The program rule initially introduced in Example 1
In this paper we were concerned with the optimization of fuzzy logic connectives
whose artificial, inefficient definitions could have been automatically produced by
previous transformation processes applied on fuzzy MALP programs. Our technique,
inspired by rewriting-based unfolding, takes profit from clear precedents in pure func-
tional programming. In this paper we have focused on the optimization of the proper
unfolding process (initially presented in [29]) by making use of dependency graphs
in order to decide the ordering in which several connective calls must be unfolded
inside a concrete connective definition. For the near future, we plan to implement our
technique inside the fuzzy logic programming environment FLOPER (visit http://
dectau.uclm.es/floper/) we have designed for developing applications coded with the
MALP language [1, 3, 40].
References
1. Almendros-Jiménez, J.M., Bofill, M., Luna Tedesqui, A., Moreno, G., Vázquez, C., Villaret,
M.: Fuzzy xpath for the automatic search of fuzzy formulae models. In: Beierle, C., Dekhtyar A.
(eds.) Scalable Uncertainty Management—9th International Conference, SUM 2015, Québec
City, QC, Canada, September 16–18, 2015. Proceedings, volume 9310 of Lecture Notes in
Computer Science, 385–398. Springer, 2015
2. Almendros-Jiménez, J.M., Luna, A., Moreno, G.: Fuzzy logic programming for implementing
a flexible xpath-based query language. Electron. Notes Theor. Comput. Sci. 282, 3–18 (2012)
3. Almendros-Jiménez, J.M., Luna, A., Moreno, G.: Fuzzy xpath through fuzzy logic program-
ming. New Generation Computing 33(2), 173–209 (2015)
4. Alpuente, M., Falaschi, M., Moreno, G., Vidal, G.: Rules + strategies for transforming lazy
functional logic programs. Theoretical Comput. Sci. Elsevier 311(1–3), 479–525 (2004)
5. Arts, T., Giesl, J.: Termination of term rewriting using dependency pairs. Theor. Comput. Sci.
236(1–2), 133–178 (2000)
6. Baader, F., Nipkow, T.: Term Rewriting and All That. Cambridge University Press, Cambridge
(1998)
7. Baldwin, J.F., Martin, T.P., Pilsworth, B.W.: Fril- Fuzzy and Evidential Reasoning in Artificial
Intelligence. Wiley, New Yark (1995)
8. Burstall, R.M., Darlington, J.: A transformation system for developing recursive programs. J.
ACM 24(1), 44–67 (1977)
9. Guerrero, J.A., Martínez, M.S., Moreno, G., Vázquez, C.: Designing lattices of truth degrees
for fuzzy logic programming environments. In: IEEE Symposium Series on Computational
Intelligence, SSCI 2015, Cape Town, South Africa, December 7–10, 2015, pp. 995–1004
IEEE, 2015
10. Guerrero, J.A., Moreno, G.: Optimizing fuzzy logic programs by unfolding, aggregation and
folding. Electron. Notes Theor. Comput. Sci. 219, 19–34 (2008)
11. Hanus, M. (ed.): Curry: An integrated functional logic language (vers. 0.8.3). Available at
http://www.curry-language.org, (2012)
12. Huet, G., Lévy, J.J.: Computations in orthogonal rewriting systems, Part I + II. In: Lassez, J.L.,
Plotkin, G.D. (eds.) Computational logic—essays in honor of alan robinson, pp. 395–443. The
MIT Press, Cambridge, MA (1992)
13. Hutton, Graham: Programming in Haskell. Cambridge University Press, Cambridge (2007)
Efficient Unfolding of Fuzzy Connectives for Multi-adjoint Logic Programs 77
14. Ishizuka, M., Kanai, N.: Prolog-ELF Incorporating Fuzzy Logic. In Joshi, A.K. (eds.) Proceed-
ings of the 9th International Joint Conference on Artificial Intelligence, IJCAI’85, 701–703.
Morgan Kaufmann, 1985
15. Julián Iranzo, P., Moreno, G., Penabad, J., Vázquez, C.: A declarative semantics for a fuzzy logic
language managing similarities and truth degrees. In: Alferes, J., Bertossi, L.E., Governatori,
G., Fodor, P., Roman, D. (eds.) Rule Technologies. Research, Tools, and Applications—10th
International Symposium, RuleML 2016, Stony Brook, NY, USA, July 6–9, 2016. Proceedings,
volume 9718 of Lecture Notes in Computer Science, 68–82. Springer, 2016
16. Julián, P., Medina, J., Morcillo, P.J., Moreno, G., Ojeda-Aciego, M.: An unfolding-based pre-
process for reinforcing thresholds in fuzzy tabulation. In: Proceedings of the 12th Interna-
tional Work-Conference on Artificial Neural Networks, IWANN’13, 647–655. Springer Verlag,
LNCS 7902, Part I, 2013
17. Julián, P., Moreno, G., Penabad, J.: On fuzzy unfolding. a multi-adjoint approach. Fuzzy Sets
Sys. 154, 16–33 (2005)
18. Julián, P., Moreno, G., Penabad, J.: Operational/interpretive unfolding of multi-adjoint logic
programs. J. Universal Comput. Sci. 12(11), 1679–1699 (2006)
19. Kifer, M., Subrahmanian, V.S.: Theory of generalized annotated logic programming and its
applications. J. Logic Program. 12, 335–367 (1992)
20. Klop, J.W., Middeldorp, A.: Sequentiality in Orthogonal Term Rewriting Systems. J. Symbol.
Comput. 161–195 (1991)
21. Lee, R.C.T.: Fuzzy logic and the resolution principle. J. ACM 19(1), 119–129 (1972)
22. Lee, C., Jones, N., Ben-Amram, A.: The size-change principle for program termination. SIG-
PLAN Not. 36(3), 81–92 (2001)
23. Li, D., Liu, D.: A fuzzy Prolog database system. Wiley, New Yark (1990)
24. Lloyd, J.W.: Foundations of Logic Programming. Springer-Verlag, Berlin (1987)
25. Medina, J., Ojeda-Aciego, M., Vojtáš, P.: Multi-adjoint logic programming with continuous
semantics. In: Proc of Logic Programming and Non-Monotonic Reasoning, LPNMR’01, vol.
2173, pp. 351–364. Springer-Verlag, LNAI, (2001)
26. Medina, J., Ojeda-Aciego, M., Vojtáš P.: A procedural semantics for multi-adjoint logic pro-
graming. In: Progress in Artificial Intelligence, EPIA’01, vol. 2258(1), pp. 290–297 Springer-
Verlag, LNAI, (2001)
27. Medina, J., Ojeda-Aciego, M., Vojtáš, P.: Similarity-based Unification: a multi-adjoint
approach. Fuzzy Sets Sys. 146, 43–62 (2004)
28. Morcillo, P.J., Moreno, G., Penabad, J., Vázquez, C.: Fuzzy Computed Answers Collecting
Proof Information. In: Cabestany, J., et al. (eds.) Advances in Computational Intelligence—
Proceedings of the 11th International Work-Conference on Artificial Neural Networks, IWANN
2011, 445–452. Springer Verlag, LNCS 6692, 2011
29. Morcillo, P.J., Moreno, G.: Improving multi-adjoint logic programs by unfolding fuzzy con-
nective definitions. In: Proceedings of the 13th International Work-Conference on Artificial
Neural Networks, IWANN 2015 Mallorca, Spain, June 10–12, pp. 511–524. Springer Verlag,
LNCS 9094, Part I, 2015
30. Morcillo, P.J., Moreno, G.: On cost estimations for executing fuzzy logic programs. In Arabnia,
H.R. et al. (ed.) Proceedings of the 11th International Conference on Artificial Intelligence,
ICAI 2009, July 13–16, 2009, Las Vegas (Nevada), USA
31. Morcillo, P.J., Moreno, G. Penabad, J., Vázquez, C.: A Practical Management of Fuzzy Truth
Degrees using FLOPER. In: Dean, M., et al. (eds.) Proceedings of 4nd Intl Symposium on
Rule Interchange and Applications, RuleML’10. Washington, USA, October 21–23, 20–34.
Springer Verlag, LNCS 6403, 2010
32. Morcillo, P.J. Moreno, G.: Modeling interpretive steps in fuzzy logic computations. In: Di Gesù
V. et al. (eds) Proceedings of the 8th International Workshop on Fuzzy Logic and Applications,
WILF 2009. Palermo, Italy, June 9-12, 44–51. Springer Verlag, LNAI 5571, 2009
33. Moreno, G., Penabad, J., Vidal. G.: Tuning fuzzy logic programs with symbolic execution.
CoRR, abs/1608.04688, 2016
78 P. J. Morcillo and G. Moreno
34. Moreno, G.: Building a fuzzy transformation system. In: Wiedermann, J. et al. (eds.) Proceed-
ings of the 32nd Conference on Current Trends in Theory and Practice of Computer Science,
SOFSEM’2006. Merin, Czech Republic, January 21–27, 409–418. Springer Verlag, LNCS
3831, 2006
35. Moreno, Ginés: A narrowing-based instantiation rule for rewriting-based fold/unfold transfor-
mations. Electr. Notes Theor. Comput. Sci. 86(3), 144–167 (2003)
36. Moreno, G., Vázquez, C.: Fuzzy logic programming in action with floper. J. Software Eng.
Appl. 7, 273–298 (2014)
37. Rodríguez-Artalejo, M., Romero-Díaz, C.: Quantitative logic programming revisited. In: Pro-
ceedings of 9th Functional and Logic Programming Symposium, FLOPS’08, 272–288. LNCS
4989, Springer Verlag, 2008
38. Straccia, U.: Managing uncertainty and vagueness in description logics, logic programs and
description logic programs. In: Reasoning Web, 4th International Summer School, Tutorial
Lectures, number 5224 in Lecture Notes in Computer Science, 54–103. Springer Verlag, 2008
39. Tamaki, H., Sato, T.: Unfold/Fold Transformations of Logic Programs. In: Tärnlund, S. (ed.)
Proceedings of Second International Conference on Logic Programming, 127–139, 1984
40. Vázquez, C., Moreno, G.. Tomás, L., Tordsson, J.: A cloud scheduler assisted by a fuzzy
affinity-aware engine. In: Yazici, A., Pal, N.R., Kaymak, U., Martin, T., Ishibuchi, H., Lin, C.-
T., Sousa, J.M.C., Tütmez, B. (eds.) 2015 IEEE International Conference on Fuzzy Systems,
FUZZ-IEEE 2015, Istanbul, Turkey, August 2–5, 2015, 1–8. IEEE, 2015
On Fuzzy Generalizations of Concept
Lattices
1 Introduction
The study of structures and mappings which allow to analyze the data in various
forms is a challenging task. In this way, the first attempts to interpret the lattice
theory as concretely as possible and to promote the better communication between
lattice theorists and potential users of lattice theory represent the inception for data
analysis taking into account the binary relations on the objects and attributes sets
[94]. Since the concept hierarchies play an important role here, the term of formal
concept analysis has been adopted for this reasoning. Briefly, formal concept anal-
ysis scrutinizes an object-attribute block of relational data in bivalent form and the
complex foundations were built in [45].
For the inception of formal concept analysis, very influential texts arise in [8, 9].
The efficient selection of relevant formal concepts is an interesting and important
issue for investigation and several studies have focused on this scalability question
in formal concept analysis. The stability index of Kuznetsov [67] represents the pro-
portion of subsets of attributes of a given concept whose closure is equal to the extent
of this concept (in an extensional formulation). A high stability index signalizes that
extent does not disappear if the extent of some of its attributes is modified. It helps to
isolate concepts that appear because of noisy objects in [52]. The complete restoring
of the original concept lattice is achieved by stability index in combination with two
other indices. The phenomenon of the basic level of concepts is advocated to select
important formal concepts in Bělohlávek et al. [22]. Five quantitative approaches on
the basic level of concepts and their metrics are comparatively analyzed in [23]. The
approaches on selecting of the formal concepts and simplifying the concept lattices
are examined by [18, 35, 40, 48–50, 66, 68], as well. A large survey on models and
techniques in knowledge processing based on formal concept analysis is given by
[84].
The applications of formal concept analysis which focus on the text retrieval and
text mining were published in the monograph [34]. The state of art of formal concept
analysis and its applications in linguistics, lexical databases, rule mining (iceberg
concept lattice), knowledge management, software engineering, object-engineering,
software development and data analysis are covered in the special LNAI 3626 Vol-
ume devoted to Formal concept analysis—Foundations and Applications and edited
by Berhard Ganter, Gerd Stumme and Rudolf Wille in 2005. An extensive overview
of applications of formal concept analysis in the various application domains includ-
ing software mining, web analytics, medicine, biology and chemistry data is given
by [83]. In addition, we mention the analysis how students learn to program [87], the
techniques for analyzing and improving integrated care pathways [82], evaluation
of questionnaires [20], or the morphological image and signal processing [2]. Our
applications aim at finding a set of representative symptoms for the disease [51],
clustering in a social network area [61] or exploring the educational tasks and objec-
tives of teachers who give the lessons of programming fundamentals [3]. We have
also applied formal concept analysis to explore the elements of MIT App Inventor 2
programming language [3].
Up to date, we remind that some other extensions called relational concept anal-
ysis, logical concept analysis, triadic concept analysis, temporal concept analysis,
rough formal concept analysis and pattern structures are sound and available for
thorough study and their application potential.
Conceptual scaling [45] and pattern structures [44] offer the possibility to process
the many-valued formal contexts. In this many-valued direction, the truth degrees
from Fuzzy logic were advocated by many researchers in effort to promote the
representation and interpretation of data in many-valued form. In this paper, we
recall early fuzzy extensions in formal concept analysis (Sect. 2), then we stress
the representation of fuzzy conjunctions (Sect. 3) in the various generalizations. We
start with the common generalized platform (Sect. 3.1), several novel heterogeneous
approaches (Sect. 3.2) and formal context of higher order (Sect. 3.3). The last one
On Fuzzy Generalizations of Concept Lattices 81
is based on the external formal contexts, harnessing the early fuzzy extensions to
the novel frameworks, and allows to explore the relationships between the formal
contexts by using the bonds in their L-fuzzy generalization.
The statements that people use to communicate facts about the world are usually not
bivalent. The truth of such statements is a matter of degree, rather than being only
true or false. Fuzzy logic and fuzzy set theory are frameworks which extend formal
concept analysis in various independent ways. Here, we recall the basic definitions
of fuzzy formal context. The structures of partially ordered set, complete lattice or
residuated lattice are applied here to represent data.
Definition 1 Consider two nonempty sets A and B, an ordered set of truth degrees
T and a mapping R such that R : A × B −→ T . Then, the triple A, B, R is called
a (T )-fuzzy formal context. The elements of the sets A and B are called attributes
and objects, respectively. The mapping R is called a (T )-fuzzy incidence relation.
In the definition of (T )-fuzzy formal context, we often take the interval T = [0, 1],
because it is a frequent scale of truth degrees in many applications. For such replace-
ment, the shortened notion of fuzzy formal context has been adopted. Analogously,
we can consider L-fuzzy formal context, or P-fuzzy formal context, having replaced
the interval [0, 1] by the algebraic structures of complete residuated lattice L, par-
tially ordered set P or other plausible scale of truth degrees.
Definition 3 Let A be a nonempty set and let T be an ordered set of truth degrees.
The mapping f : A −→ T is called a fuzzy membership function (also called a T -
fuzzy set or a fuzzy set). We denote a set of all fuzzy membership functions A −→ T
by T A .
and
( ( f ))(b) = f (b) = ( f (a) → R(a, b)). (2)
a∈A
are called L-fuzzy concept-forming operators (or L-fuzzy derivation operators, short-
ened L-fuzzy polars).
The set of pairs
L-FCL( A, B, R) = { f, g : f = g, g = f} (3)
is called the set of all L-fuzzy formal concepts of A, B, R. The elements of the
pairs f, g ∈ L-FCL( A, B, R) are called the extents and intents of L-fuzzy formal
concepts, respectively. The set of all extents is denoted by Ext(A, B, R), the set of
all intents by Int(A, B, R).
Bělohlávek [11] proved that L-fuzzy concept-forming operators form an antitone
Galois connection. In order theory [39, 45, 78], an antitone Galois connection is given
by two opposite mappings between ordered sets. These mappings are order-inverting
and a composition of these two mappings yields two closure operators.
Definition 5 An antitone Galois connection between two ordered sets (P, ≤ P ) and
(Q, ≤ Q ) is a pair g, f of mappings g : P → Q, f : Q → P such that for all p ∈ P
and q ∈ Q
g( p) ≤ Q q i f andonlyi f p ≥ P f (q).
Other useful properties of L-fuzzy formal concepts have been intensively inves-
tigated since their first inception in [11]. The structure of L-fuzzy concept lattice is
given by the set of all L-fuzzy formal concepts equipped with a crisp order here.
In the research line, the papers [12–14] are oriented to fuzzy Galois connections
and L-fuzzy concept lattices which are enhanced by a fuzzy order of L-fuzzy formal
On Fuzzy Generalizations of Concept Lattices 83
concepts. More extensive results about these issues have been proliferated in [15, 17,
19, 24–26] and the more thorough findings can be found in the recently published
literature.
The extents and intents of L-fuzzy formal concepts are both represented by fuzzy
membership functions. In [11], Bělohlávek explains the examples of empirically
empty fuzzy extents, partially empty fuzzy extents (which can be interpreted appar-
ently natural) and full crisp extents in effort to interpret L-fuzzy concepts given by
planets and their attributes.
Full crisp extents with the maintenance of fuzzy intents became the origin of our
approach which was elaborated theoretically [56], as well as practically in our special
oriented applications for management of teams [60, 61].
Definition 6 Let X ⊆ B and ↑: P(B) −→ [0, 1] A . Then, ↑ is a mapping that assigns
to every crisp set X of objects a fuzzy membership function X ↑ of attributes, such
that a value in a point a ∈ A is:
X ↑ (a) = R(a, b). (4)
b∈X
{b1 ,b2}, 0.8, 0.7, 0.7 {b1 ,b4 }, 0.8, 0.6, 0.8
∅, 1, 1, 1
the clusters of objects due to their minimal distance function which is specially
defined. The metric properties of this distance function are proved in [56] and the
efficient algorithm is presented and illustrated. Recently, the Gaussian probabilistic
index [5] was introduced in effort to describe the stability of the extents due to random
fluctuation of values in [0, 1]-fuzzy formal context.
The applications of one-sided approach in a social network area, specially in a
secondary school class in Slovakia [60, 61] help to understand the structure of the
managed team or real school classes. The school data are collected persistently and
the mentioned methods for selection of relevant clusters are applied to serve the
complex information for the class teacher about relationships between schoolmates.
On Fuzzy Generalizations of Concept Lattices 85
Last but not least, the fuzzy subsets in one coordinate and the crisp subsets in the
other coordinate of a formal concept are independently investigated in [21, 27], as
well. Authors of [21] accentuate the crisply generated fuzzy concepts in effort to deal
with the problem of a possibly large number of concepts in an L-fuzzy approach.
The relationships between the objects and attributes are represented by a fuzzy
incidence relation. Fuzzy conjunctions describe the relationships between the truth
degrees of objects and the truth degrees of attributes. In L-fuzzy approach, the fuzzy
conjunctions L × L → L usually describe these relationships.
The inevitability to represent a diverse form of fuzzy conjunctions in a fuzzy for-
mal context leads to the novel definitions of the fuzzy formal contexts in a broader
perspective. Therefore, the notions of generalized formal context, multi-adjoint for-
mal context, heterogeneous formal context, connectional formal context and formal
context of second-order were recently introduced. Each of them allows to represent
data in a specific form and offers the interpretation on the examples from the real
world. The list of novel fuzzy approaches follows in Table 2.
The understanding the way how the fuzzy conjunctions are represented is an
important issue for potential applications. In the following subsections, we present
the cornerstone of generalized, heterogeneous and second-order version of a formal
context.
L L L
L 0 0.5 1 L 0 0.5 1 L 0 0.5 1
0 0 0 0 0 0 0 0 0 0 0 0
0.5 0 0.5 0.5 0.5 0 0 0.5 0.5 0 0.25 0.5
1 0 0.5 1 1 0 0.5 1 1 0 0.5 1
connected. In L-fuzzy approach, the frequently used formulas for fuzzy conjunction
⊗ : L × L −→ L are:
– Gödel fuzzy conjunction given by a ⊗ b = min(a, b),
– Łukasiewicz fuzzy conjunction given by a ⊗ b = max(0, a + b − 1),
– product fuzzy conjunction given by a ⊗ b = a · b.
The Gödel fuzzy conjunction points to a smaller number of formal concepts in
comparison with the Łukasiewicz fuzzy conjunction. Moreover, a smaller fluctuation
of fuzzy membership function usually causes a smaller fluctuation of its closure com-
puted by the Łukasiewicz fuzzy conjunction in comparison with the Gödel structure.
The product fuzzy conjuntion can be discretized to compute the formal concepts in
L-fuzzy approach. However, notice that the Gödel, Łukasiewicz and product fuzzy
conjunctions are commutative, monotone and left-continuous operations. The exam-
ples of Gödel, Łukasiewicz and product fuzzy conjunctions are shown in Fig. 5.
The first prospects to non-commutativity of fuzzy conjunctions L × L −→ L is
given by Georgescu and Popescu in [47]. This non-commutative extension requires
the definition of L-fuzzy concept with one extent and two intents.
Analogously in [58], our motivation is based on a more generalized formula-
tion of fuzzy conjunctions in comparison with L-fuzzy formal context (also for
non-commutative cases). The aim is to formulate fuzzy conjunctions by three-
sorted structures of truth degrees which are not necessarily numerical, see for
instance ({s, t, u}, s ≤ t ≤ u) or ({⊥, }, ⊥ ≤ ). Another possibility is to con-
sider not linearly ordered set of truth degrees, e. g. ({0, s, t, 1}, 0 ≤ s ≤ 1, 0 ≤ t ≤
1, s is incomparable with t). We aim at preserving the one extent and one intent in a
formal concept.
Definition 9 Let A, B, R be a P-fuzzy formal context for a poset (P, P ). Let
(C, C ) and (D, D ) be the complete lattices.
Let • be an operation (a fuzzy conjunction) such that • is from C × D to P and
is monotone and left-continuous in both arguments, that is:
(1a) c1 C c2 implies c1 • d P c2 • d for all c1 , c2 ∈ C and d ∈ D.
(1b) d1 D d2 implies c • d1 P c • d2 for all c ∈ C and d1 , d2 ∈D.
(2a) If c • d P p for d ∈ D, p ∈ P and for all c ∈ S ⊆ C, then S• d P p.
(2b) If c • d P p for c ∈ C, p ∈ P and for all d ∈ T ⊆ D, then c • T P p.
Then, the tuple A, B, P, R, C, D, • is called a generalized formal context.
On Fuzzy Generalizations of Concept Lattices 87
The left part of Fig. 6 indicates non-commutative fuzzy conjunction. The right
part expresses a fuzzy conjunction in a generalized formal context.
( G (g))(a) = {c ∈ C : (∀b ∈ B)c • g(b) P R(a, b)},
( G( f ))(b) = {d ∈ D : (∀a ∈ A) f (a) • d P R(a, b)}
experiments, they claim that a stronger hedge lead to a smaller L-fuzzy concept
lattice. The close relationship between the generalized concept lattices and L-fuzzy
concept lattices constrained by truth-stressing hedges is explored in [59]. The foun-
dations of isotone fuzzy Galois connections with a truth-stressing hedge and a truth-
depressing hedge were developed by Konečný [53] and the reduction of trilattices
with truth-stressing hedges is described by Konečný and Osička in [55].
The reasoning about heterogeneous data in a formal context appears in [36–38, 70,
72, 79, 86]. These works inspired us to define the heterogeneous extension based on
the full diversification of structures, whereby the construction of concept lattices is
still sound and the proposed extension covers the well-known approaches.
Particularly, the main idea is based on a diversification of all data structures that
can be diversified within a formal context. We use different sets of truth degrees across
a set of objects, different sets of truth degrees across a set of attributes and different
sets of truth degrees across the object-attribute table fields. In addition, for each
object-attribute pair, one can formulate the original fuzzy conjunctions according
to the three-sorted sets of truth degrees of each particular object, attribute and type
of values of corresponding table field. The full definition of heterogeneous formal
context follows.
Definition 11 Consider the set of attributes A and the set of objects B. Let P =
((Pa,b , Pa,b ) : a ∈ A, b ∈ B) be a system of posets and let R be a function from
A × B such that R(a, b) ∈ Pa,b for all a ∈ A, b ∈ B. Let C = ((Ca , Ca ) : a ∈ A)
and D = ((Db , Db ) : b ∈ B) be systems of complete lattices.
Let = (•a,b : a ∈ A, b ∈ B) be a system of operations such that •a,b is from
Ca × Db to Pa,b and is monotone and left-continuous in both arguments, that is:
(1a) c1 c2 implies c1 •a,b d c2 •a,b d for all c1 , c2 ∈ Ca and d ∈ Db .
(1b) d1 d2 implies c •a,b d1 c •a,b d2 for all c ∈ Ca and d1 , d2 ∈ Db.
(2a) If c •a,b d p for d ∈ Db , p ∈ Pa,b and for all c ∈ S ⊆ Ca , then S •a,b
d p.
(2b) If c •a,b d p for c ∈ Ca , p ∈ Pa,b and for all d ∈ T ⊆ Db , then c •a,b
T p.
Then, we call the tuple A, B, P, R, C, D, a heterogeneous formal context.
( H (g))(a) = {c ∈ Ca : (∀b ∈ B)c •a,b g(b) R(a, b)}.
Symmetrically, let f ∈ a∈A Ca and H : a∈A Ca −→ b∈B Db . Then, H is
a mapping that assigns to every function f a function H ( f ), such that its value in
the object b is given by
( H( f ))(b) = {d ∈ Db : (∀a ∈ A) f (a) •a,b d R(a, b)}.
and
gi , f i = H H gi , fi .
i∈I i∈I i∈I
90 L. Antoni et al.
2. For each a ∈ A and b ∈ B, let 0 Pa,b be the least element of Pa,b such that
0Ca •a,b d = c •a,b 0 Db = 0 Pa,b for all c ∈ Ca , d ∈ Db . Then, a complete lattice
L is isomorphic to HCL(A, B, P, R, C, D, , H , H , ≤) if and only if there
are mappings α : a∈A ({a} × Ca ) −→ L and β : b∈B ({b} × Db ) −→ L such
that:
(1a) α does not increase in the second argument (for a fixed first argument);
(1b) β does not decrease in the second argument (for a fixed first argument);
(2a) Rng(α) is ∧-dense in L;
(2b) Rng(β) is ∨-dense in L; and
(3) For every a ∈ A, b ∈ B, c ∈ Ca and d ∈ Db ,
Other inspiring approaches describe the possibility to apply formal concept anal-
ysis for heterogeneous data, as well. Pócs introduces the connectional approach in
[79, 80] which is based on the antitone Galois connections (Definition 5) between
each pair of object and attribute. Fuzzy conjunctions are replaced by the antitone
Galois connections. Note that the definition of connectional formal context does not
contain the fuzzy incidence relation.
For f ∈ a∈A Ca , the mapping G : a∈A Ca −→ b∈B Db is defined by
( G( f ))(b) = ψa,b ( f (a)).
a∈A
l m t
k s
from the properties of an antitone Galois connection that φa1 ,b1 (k) = u, φa1 ,b1 (l) = u,
φa1 ,b1 (m) = t, φa1 ,b1 (n) = t, ψa1 ,b1 (s) = n, ψa1 ,b1 (t) = n and ψa1 ,b1 (u) = l.
Medina and Ojeda-Aciego applied the idea of multi-adjointness from logic pro-
gramming [76, 77] to formal concept analysis. The multi-adjoint formal context is
characterized by the adjoints pertaining to the operations • in a common generalized
platform. The first prospects, motivating results and examples were published in [73,
75]. We recall a multi-adjoint formal context here. Two central notions are an adjoint
triple and a multi-adjoint frame.
Definition 14 Let (U, U ), (V, V ) and (P, P ) be posets. Then, the triple (•, →1
, →2 ) is called an adjoint triple (or implication triple) if • : (U × V ) −→ P, →1 :
(V × P) −→ U and →2 : (U × P) −→ V and moreover
Definition 15 Consider the set of attributes A and the set of objects B. Let C =
((Ca , Ca ) : a ∈ A) be a system of complete lattices. Let R be a function such that
R(a, b) ∈ Ca for all a ∈ A, b ∈ B. Then, we call a tuple A, B, R, C a heterogeneous
one-sided fuzzy formal context (or shortly C-fuzzy formal context).
ized one-sided fuzzy approach has been adopted there, but we call it heterogeneous
in effort to emphasize the heterogeneous types of attributes in this extension.
An alternative approach to solve the presence of heterogeneous data in a fuzzy
formal context, taking into account the structures of idempotent semirings, can be
seen in the papers [92, 93] of Valverde-Albacete and Peláez-Moreno.
Definition 16 Consider two non-empty index sets J , I and L-fuzzy formal context
j∈J A j , i∈I Bi , R, whereby
– A j1 ∩ A j2 = ∅ for any j1 , j2 ∈ J , j1 = j2 ,
– Bi1 ∩ Bi2 = ∅ for any i 1 , i 2 ∈ I , i 1 = i 2 ,
– R : j∈J A j × i∈I Bi −→ L.
Moreover, consider two non-empty sets of L-fuzzy formal contexts (external formal
contexts) notated by
– {O j , A j , Q j : j ∈ J }, whereby D j = O j , A j , Q j ,
– {Bi , Ti , Pi : i ∈ I }, whereby Ci = Bi , Ti , Pi .
A second-order formal context is a tuple
A j , {D j ; j ∈ J }, Bi , {Ci ; i ∈ I }, R ,
j∈J i∈I
Fig. 10 Example of D1 D2
second-order formal context
o1 1 0.5 0 1 o4
o2 0.5 0.5 1 0 o5
o3 1 1 1 1 o6
t3 t2 t1 a1 a2 a3 a4
0.5 0.5 1 b1 1 0 1 1
C1 1 1 0.5 b2 0.5 0.5 1 1
1 1 0.5 b3 0.5 1 1 0.5
0 1 b4 1 1 0.5 0
C2
1 1 b5 1 1 0.5 1
t5 t4
can also consider only one external formal context for objects and only one external
formal context for attributes in a very special case. Furthermore, the extents and
intents of these external formal contexts represent the inputs for the computations of
second-order formal concepts.
The direct product of external
L-fuzzy lattices L-FCL(C1 ) × L-FCL(C2 ) × . . . ×
L-FCL(Cn ) is denoted by i∈{1,2,...,n} L-FCL(Ci ). To find the second-order formal
concepts, the second-order concept-forming operators are defined between direct
products of the two previously described sets of L-fuzzy concept lattices. The precise
form of the concept-forming operators are introduced in [62] and their formulation
suppose the equivalence functor between the categories of fuzzy Chu correspon-
dences and completely lattice L-ordered sets. For more details, see [62, 65].
Nevertheless, to compute the second-order formal concepts in a more shortened
way, we can use the additional results presented in [62]. To do that, we need to
recall that given two arbitrary sets A and B, the mapping β : B → L A is called
L-multimapping. Then, a notion of L-bond is defined as follows.
Definition 17 Let Ci = Ai , Bi , Ri for i ∈ {1, 2} be two L-fuzzy formal contexts.
An L-bond is an L-multimapping β : B1 −→ Int(C2 ) such that β t : A2 −→ Ext(C1 ),
where β t (a2 )(b1 ) = β(b1 )(a2 ) for any (b1 , a2 ) ∈ B1 × A2 . The set of all L-bonds
between L-fuzzy formal contexts C1 and C2 is denoted by L-Bonds(C1 , C2 ).
Remark 1 The set of all L-bonds between two L-fuzzy formal contexts forms the
structure of complete lattice. For the proof and the precise formulation of supremum
and infimum of L-bonds see [64]. Relationships between L-bonds and extents of
direct products of L-fuzzy contexts are drawn in [63].
The complete lattice of the nine L-bonds between D1 and C1 (from Fig. 10) can
be seen in Fig. 11. For the calculations of the extents and intents of L-fuzzy formal
contexts by Equation (1) and (2), we selected the Łukasiewicz fuzzy implications
given by a → b = min{1, 1 − a + b} for a, b ∈ L.
Having selected one particular L-bond β from the set of all L-bonds between two
L-fuzzy formal contexts, one can introduce the concept-forming operators between
On Fuzzy Generalizations of Concept Lattices 95
β7 β8
1 0.5 1 1
1 1 1 0.5
1 1 1 0.5
β4 β5 β6
0.5 0.5 1 0.5 1 1
1 1 1 0.5 0.5 0.5
1 1 1 0.5 0.5 0.5
β2 β3
0.5 0.5 1 0.5
1 0.5 0.5 0.5
1 0.5 0.5 0.5
β1
0.5 0.5
0.5 0.5
0.5 0.5
96 L. Antoni et al.
two external sets of all L-fuzzy membership functions (L B1 and L A2 ) as the following
definition states.
Definition 18 Let Ci = Ai , Bi , Ri for i ∈ {1, 2} be two L-fuzzy contexts and let
β be an arbitrary L-bond between C1 and C2 . The mappings ↑β : L B1 −→ L A2 and
↓β : L A2 −→ L B1 such that
(↑β ( f ))(a) = ( f (b) → β(b)(a)) (6)
b∈B1
and
(↓β (g))(b) = (g(a) → β(b)(a)) (7)
a∈A2
K= A j , {D j : j ∈ J }, Bi , {Ci : i ∈ I }, R
j∈J i∈I
=
K Aj, Bi , ρ ji ,
j∈J i∈I ( j,i)∈J ×I
ρ ji = {β ∈ L-Bonds(D j , Ci ) : (∀(a j , bi ) ∈ A j × Bi )β(a j )(bi ) ≥ R ji (a j , bi )}
are isomorphic.
be an L-fuzzy formal context. Then, the concept lattices of K and K
Proof The proof is presented in [62].
In an L-fuzzy formal context K given by Theorem 2, the subrelation R ji of a
second-order formal context K is replaced by an L-bond ρ ji for each pair ( j, i) ∈
J × I . The bond ρ ji is constructed as the closest bond with respect to the subrelation
R ji . For more details and examples, see [62]. In the second-order formal context given
by Fig. 10, we have that ρ1,1 = β7 , whereby β7 is described in Fig. 11.
We have described the way how to find the second-order formal concepts. The
heterogeneous formal contexts from Definition 11 can be seen in terms of the second-
order formal contexts as the following translation indicates:
On Fuzzy Generalizations of Concept Lattices 97
4 Conclusion
The information hidden within data can help to solve the many pending issues within
community, enterprise or science. Turning of data into knowledge and wisdom is
beneficial and necessary, considering either the simple computing in the spread-
sheet calculators or various methods of data analysis which are more complex. Data
collecting, preprocessing, reduction, visualization and dependencies exploration are
important parts of the scientific research, as well.
98 L. Antoni et al.
heterogeneous connectional
( P ócs)
multi-adjoint generalized
(Medina, Ojeda-Aciego) (Krajči)
one-sided
( B ěloh lá vek et a l., Ya h ia et a l., K r a jči)
classical
(Gan ter and Wille)
In conclusion, we are convinced that the area of concept lattices plays an important
role in the research of both theoreticians and practitioners and thus, the research from
a broader perspective is still beneficial.
Acknowledgements We thank the anonymous reviewers for their careful reading of our manuscript
and their many fruitful comments and suggestions. This work was partially supported by the Sci-
entific Grant Agency of the Ministry of Education of Slovak Republic and the Slovak Academy
of Sciences under the contract VEGA 1/0073/15 and by the Slovak Research and Development
Agency under the contract No. APVV–15–0091. This work was partially supported by the Agency
of the Ministry of Education, Science, Research and Sport of the Slovak Republic for the Structural
Funds of EU under the project Center of knowledge and information systems in Košice, CeZIS,
ITMS: 26220220158.
References
1. Alcalde, C., Burusco, A., Fuentes-González, R., Zubia, I.: The use of linguistic variables and
fuzzy propositions in the L-fuzzy concept theory. Comput. Math. Appl. 62(8), 3111–3122
(2011)
2. Alcalde, C., Burusco, A., Fuentes-González, R.: Application of the L-fuzzy concept analysis
in the morphological image and signal processing. Ann. Math. Artif. Intell. 72(1–2), 115–128
(2014)
3. Antoni, L., Guniš, J., Krajči, S., Krídlo, O., Šnajder, L.: The educational tasks and objectives
system within a formal context. In: Bertet, K., Rudolph, S. (eds.) CLA 2014. CEUR Workshop
Proceedings, vol. 1252, pp. 35–46. CEUR-WS.org (2014)
4. Antoni, L., Krajči, S., Krídlo, O.: On different types of heterogeneous formal contexts. In: Pasi,
G., Montero, J., Ciucci, D. (eds.) Proceedings of the 8th Conference of the European Society
for Fuzzy Logic and Technology, pp. 302–309. Atlantis Press (2013)
5. Antoni, L., Krajči, S., Krídlo, O.: Randomized fuzzy formal contexts and relevance of one-
sided concepts. In: Baixeries, J., Sacarea, Ch. (eds.) ICFCA 2015. LNCS (LNAI), vol. 9113.
Springer, Heidelberg (2015)
6. Antoni, L., Krajči, S., Krídlo, O., Macek, B., Pisková, L.: Relationship between two FCA
approaches on heterogeneous formal contexts. In: Szathmary, L., Priss, U. (eds.) CLA 2012.
CEUR Workshop Proceedings, vol. 972, pp. 93–102. CEUR-WS.org (2012)
7. Antoni, L., Krajči, S., Krídlo, O., Macek, B., Pisková, L.: On heterogeneous formal contexts.
Fuzzy Sets Syst. 234, 22–33 (2014)
8. Arnauld, A., Nicole, P.: Logic, or the Art of Thinking. Cambridge University Press, Cambridge
(1996)
9. Barbut, M., Monjardet, B.: Ordre et classification. Algbre et combinatoire. Collection Hachette
Universit, Librairie Hachette (1970)
10. Bartl, E.: Minimal solutions of generalized fuzzy relational equations: probabilistic algorithm
based on greedy approach. Fuzzy Sets Syst. 260, 25–42 (2015)
11. Bělohlávek, R.: Lattices generated by binary fuzzy relations. Tatra Mt. Math. Publ. 16, 11–19
(1999)
12. Bělohlávek, R.: Fuzzy galois connections. Math. Logic Q. 45(4), 497–504 (1999)
13. Bělohlávek, R.: Reduction and a simple proof of characterization of fuzzy concept lattices.
Fundam. Informaticae 46(4), 277–285 (2001)
14. Bělohlávek, R.: Concept lattices and order in fuzzy logic. Ann. Pure Appl. Logic 128, 277–298
(2004)
15. Bělohlávek, R.: What is fuzzy concept lattice? II. In: Kuznetsov, S.O., Slezak, D., Hepting,
D.H., Mirkin, B.G. (eds.) Rough Sets, Fuzzy Sets, Data Mining and Granular Computing.
LNCS, vol. 6743, pp. 19–26. Springer, Heidelberg (2011)
100 L. Antoni et al.
16. Bělohlávek, R.: Sup-t-norm and inf-residuum are one type of relational product: Unifying
framework and consequences. Fuzzy Sets Syst. 197, 45–58 (2012)
17. Bělohlávek, R.: Ordinally equivalent data: a measurement-theoretic look at formal concept
analysis of fuzzy attributes. Int. J. Approx. Reason. 54(9), 1496–1506 (2013)
18. Bělohlávek, R., De Baets, B., Konečný, J.: Granularity of attributes in formal concept analysis.
Inf. Sci. 260, 149–170 (2014)
19. Bělohlávek, R., Klir, G.J.: Concepts and Fuzzy Logic. MIT Press, Cambridge (2011)
20. Bělohlávek, R., Sigmund, E., Zacpal, J.: Evaluation of IPAQ questionnaires supported by formal
concept analysis. Inf. Sci. 181(10), 1774–1786 (2011)
21. Bělohlávek, R., Sklenář, V., Zacpal, J.: Crisply generated fuzzy concepts. In: Ganter, B., Godin,
R. (eds.) ICFCA 2005. LNCS (LNAI), vol. 3403, pp. 269–284. Springer, Heidelberg (2005)
22. Bělohlávek, R., Trnečka, M.: Basic level of concepts in formal concept analysis. In: Domenach,
F., Ignatov, D., Poelmans, J. (eds.) ICFCA 2012. LNCS (LNAI), vol. 7278, pp. 28–44. Springer,
Heidelberg (2012)
23. Bělohlávek, R., Trnečka, M.: Basic Level in Formal Concept Analysis: Interesting Concepts
and Psychological Ramifications. In: Rossi, F. (eds.) IJCAI 2013, pp. 1233–1239. AAAI Press
(2013)
24. Bělohlávek, R., Vychodil, V.: What is fuzzy concept lattice?. In: Bělohlávek, R., Snášel, V.
(eds.) CLA 2005. CEUR Workshop Proceedings, vol. 162, pp. 34–45. CEUR-WS.org (2005)
25. Bělohlávek, R., Vychodil, V.: Fuzzy concept lattices constrained by hedges. J. Adv. Comput.
Intell. Intell. Inform. 11(6), 536–545 (2007)
26. Bělohlávek, R., Vychodil, V.: Formal concept analysis and linguistic hedges. Int. J. Gen. Syst.
41(5), 503–532 (2012)
27. Ben Yahia, S., Jaoua, A.,: Discovering knowledge from fuzzy concept lattice. In: Kandel, A.,
Last, M., Bunke, H. (eds.) Data Mining and Computational Intelligence. Studies in Fuzziness
and Soft Computing, vol. 68, pp. 167–190. Physica-Verlag, Heidelberg (2001)
28. Burusco, A., Fuentes-González, R.: The study of L-fuzzy concept lattice. Mathw. Soft Comput.
3, 209–218 (1994)
29. Butka, P., Pócs, J.: Generalization of one-sided concept lattices. Comput. Inform. 32(2), 355–
370 (2013)
30. Butka, P., Pócs, J., Pócsová, J.: Representation of fuzzy concept lattices in the framework of
classical FCA. J. Appl. Math. 2013, Article ID 236725, 7 pages. (2013)
31. Butka, P., Pócs, J., Pócsová, J.: On equivalence of conceptual scaling and generalized one-sided
concept lattices. Inf. Sci. 259, 57–70 (2014)
32. Butka, P., Pócs, J., Pócsová, J.: Reduction of concepts from generalized one-sided concept
lattice based on subsets quality measure. In: Zgrzywa, A., Choros, A., Sieminski, A. (eds.) New
Research in Multimedia and Internet Systems. Advances in Intelligent Systems and Computing,
vol. 314, pp. 101–111. Springer (2015)
33. Cabrera, I.P., Cordero, P., Gutiérez, G., Martinez, J., Ojeda-Aciego, M.: On residuation in
multilattices: filters, congruences, and homorphisms. Fuzzy Sets Syst. 234, 1–21 (2014)
34. Carpineto, C., Romano, G.: Concept Data Analysis. Wiley, Theory and Applications. J (2004)
35. Cellier, P., Ferré, S., Ridoux, O., Ducassé, M.: A parameterized algorithm to explore formal
contexts with a taxonomy. Int. J. Found. of Comput. Sci. 2, 319–343 (2008)
36. Cornejo, M.E., Medina, J., Ramírez-Poussa, E.: Attribute and size reduction mechanisms in
multi-adjoint concept lattices. J. Comput. Appl. Math. 318, 388–402 (2017)
37. Cornejo, M.E., Medina, J., Ramírez-Poussa, E.: A comparative study of adjoint triples. Fuzzy
Sets Syst. 211, 1–14 (2013)
38. Cornejo, M.E., Medina, J., Ramírez-Poussa, E.: Adjoint negations, more than residuated nega-
tions. Inf. Sci. 345, 355–371 (2016)
39. Davey, B.A., Priestley, H.A.: Introduction to Lattices and Order. Cambridge University Press,
Cambridge (2002)
40. Dias, S.M., Vieira, N.J.: Concept lattices reduction: Definition, analysis and classification. In:
Expert Syst. Appl., https://doi.org/10.1016/j.eswa.2015.04.044
On Fuzzy Generalizations of Concept Lattices 101
41. Diáz-Moreno, J.C., Medina, J., Ojeda-Aciego, M.: On basic conditions to generate multi-adjoint
concept lattices via Galois connections. Int. J. Gen. Syst. 43(2), 149–161 (2014)
42. Dubois, D., Prade, H.: Possibility theory and formal concept analysis: Characterizing indepen-
dent sub-contexts. Fuzzy Sets Syst. 196, 4–16 (2012)
43. Frič, R., Papčo, M.: A categorical approach to probability theory. Stud. Logica 94(2), 215–230
(2010)
44. Ganter, B., Kuznetsov, S.: Pattern structures and their projections. In: Delugach, H.S., Stumme,
G. (eds.) ICCS 2001. LNCS, vol. 2120, pp. 129–142. Springer, Heidelberg (2001)
45. Ganter, B., Wille, R.: Formal Concept Analysis: Mathematical Foundation. Springer, Heidel-
berg (1999)
46. Garciá-Pardo, F., Cabrera, I.P., Cordero, P., Ojeda-Aciego, M., Rodríguez-Sanchez, F.J.: On
the definition of suitable orderings to generate adjunctions over an unstructured codomain. Inf.
Sci. 286, 173–187 (2014)
47. Georgescu, G., Popescu, A.: Concept lattices and similarity in non-commutative fuzzy logic.
Fundamenta Informaticae 53(1), 23–54 (2002)
48. Halaš, R., Mesiar, R., Pócs, J.: Description of sup- and inf-preserving aggregation functions
via families of clusters in data tables, Inf. Sci., to appear. https://doi.org/10.1016/j.ins.2017.
02.060
49. Halaš, R., Pócs, J.: Generalized one-sided concept lattices with attribute preferences. Inf. Sci.
303, 50–60 (2015)
50. Kardoš, F., Pócs, J., Pócsova, J.: On concept reduction based on some graph properties. Knowl.-
Based Syst. 93, 67–74 (2016)
51. Kiseliová, T., Krajči, S.: Generation of representative symptoms based on fuzzy concept lattices.
Adv. Soft Computing 33, 349–354 (2005)
52. Klimushkin, M., Obiedkov, S., Roth, C.: Approaches to the selection of relevant concepts in
the case of noisy data. In: Kwuida, L., Sertkaya, B. (eds.) ICFCA 2010. LNCS, vol. 5986, pp.
255–266. Springer, Heidelberg (2010)
53. Konečný, J.: Isotone fuzzy Galois connections with hedges. Inf. Sci. 181, 1804–1817 (2011)
54. Konečný, J., Medina, J., Ojeda-Aciego, M.: Multi-adjoint concept lattices with heterogeneous
conjunctors and hedges. Ann. Math. Artif. Intell. 72(1–2), 73–89 (2014)
55. Konečný, J., Osička, P.: Triadic concept lattices in the framework of aggregation structures.
Inf. Sci. 279, 512–527 (2014)
56. Krajči, S.: Cluster based efficient generation of fuzzy concepts. Neural Netw. World 13(5),
521–530 (2003)
57. Krajči, S.: The basic theorem on generalized concept lattice. In Bělohlávek R., Snášel V. (eds.)
CLA 2004. CEUR Workshop Proceedings, vol. 110, pp. 25–33. CEUR-WS.org (2004)
58. Krajči, S.: A generalized concept lattice. Log. J. IGPL 13(5), 543–550 (2005)
59. Krajči, S.: Every concept lattice with hedges is isomorphic to some generalized concept lattice.
In Bělohlávek R., Snášel V. (eds.) CLA 2005. CEUR Workshop Proceedings, vol. 162, pp. 1–9.
CEUR-WS.org (2005)
60. Krajči, S.: Social Network and Formal Concept Analysis. In: Pedrycz, W., Chen, S.M. (eds.)
Social Networks: A framework of Computational Intelligence. Studies in Computational Intel-
ligence, vol 526, pp. 41–62. Springer (2014)
61. Krajči, S., Krajčiová, J.: Social network and one-sided fuzzy concept lattices. In: Spyropou-
los, C. (eds.) Fuzz-IEEE 2007. Proceedings of the IEEE International Conference on Fuzzy
Systems, pp. 1–6. IEEE Press (2007)
62. Krídlo, O., Krajči, S., Antoni, L.: Formal concept analysis of higher order. Int. J. Gen. Syst.
45(2), 116–134 (2016)
63. Krídlo, O., Krajči, S., Ojeda-Aciego, M.: The category of L-Chu correspondences and the
structure of L-bonds. Fund. Inform. 115(4), 297–325 (2012)
64. Krídlo, O., Ojeda-Aciego, M.: On L-fuzzy Chu correspondences. Int. J. Comput. Math. 88(9),
1808–1818 (2011)
65. Krídlo, O., Ojeda-Aciego, M.: Revising the link between L-Chu correspondences and com-
pletely lattice L-ordered sets. Ann. Math. Artif. Intell. 72(1–2), 91–113 (2014)
102 L. Antoni et al.
66. Krupka, M.: On complexity reduction of concept lattices: three counterexamples. Inf. Retr.
15(2), 151–156 (2012)
67. Kuznetsov, S.O.: On stability of a formal concept. Ann. Math. Artif. Intell. 49, 101–115 (2007)
68. Kwuida, L., Missaoui, R., Ben Amor, B., Boumedjout, L., Vaillancourt, J.: Restrictions on
concept lattices for pattern management. In: Kryszkiewicz, M., Obiedkov, S. (eds.) CLA 2010.
CEUR Workshop Proceedings, vol. 672, pp. 235–246. CEUR-WS.org (2010)
69. Medina, J., Ojeda-Aciego, M.: Multi-adjoint t-concept lattices. Inf. Sci. 180(5), 712–725 (2010)
70. Medina, J., Ojeda-Aciego, M.: On multi-adjoint concept lattices based on heterogeneous con-
junctors. Fuzzy Sets Syst. 208, 95–110 (2012)
71. Medina, J., Ojeda-Aciego, M.: Dual multi-adjoint concept lattices. Inf. Sci. 225, 47–54 (2013)
72. Medina, J., Ojeda-Aciego, M., Pócs, J., Ramiréz-Poussa, E.: On the Dedekind-MacNeille
completion and formal concept analysis based on multilattices. Fuzzy Sets Syst. 303, 1–20
(2016)
73. Medina, J., Ojeda-Aciego, M., Ruiz-Calviño, J.: On multi-adjoint concept lattices: definition
and representation theorem. Lect. Notes Artif. Intell 4390, 197–209 (2007)
74. Medina, J., Ojeda-Aciego, M., Ruiz-Calviño, J.: Relating generalized concept lattices and
concept lattices for non-commutative conjunctors. Appl. Math. Lett. 21, 1296–1300 (2008)
75. Medina, J., Ojeda-Aciego, M., Ruiz-Calviño, J.: Formal concept analysis via multi-adjoint
concept lattices. Fuzzy Sets Syst. 160(2), 130–144 (2009)
76. Medina, J., Ojeda-Aciego, M., Valverde, A., Vojtáš, P.: Towards biresiduated multi-adjoint logic
programming. In: Conejo, R., Urretavizcaya, M., Pérez-de-la-Cruz, J.-L. (eds.) CAEPIA-TTIA
2003. LNCS (LNAI), vol. 3040, pp. 608–617. Springer, Heidelberg (2004)
77. Medina, J., Ojeda-Aciego, M., Vojtáš, P.: Similarity-based unification: a multi-adjoint approach.
Fuzzy Sets Syst. 146, 43–62 (2004)
78. Ore, Ø.: Galois connexions. Trans. Am. Mathe. Soc. 55, 493–513 (1944)
79. Pócs, J.: Note on generating fuzzy concept lattices via Galois connections. Inf. Sci. 185(1),
128–136 (2012)
80. Pócs, J.: On possible generalization of fuzzy concept lattices using dually isomorphic retracts.
Inf. Sci. 210, 89–98 (2012)
81. Pócs, J., Pócsová, J.: On some general aspects of forming fuzzy concept lattices. Appl. Math.
Sci. 7(112), 5559–5605 (2013)
82. Poelmans, J., Dedene, G., Verheyden, G., Van der Mussele, H., Viaene, S., Peters, E.: Combining
business process and data discovery techniques for analyzing and improving integrated care
pathways. In: Perner, P. (ed.) ICDM 2010, pp. 505–517. Springer-Verlag, Berlin Heidelberg
(2010)
83. Poelmans, J., Ignatov, D.I., Kuznetsov, S.O., Dedene, G.: Formal concept analysis in knowledge
processing: a survey on applications. Expert Syst. Appl. 40(16), 6538–6560 (2013)
84. Poelmans, J., Kuznetsov, S.O., Ignatov, D.I., Dedene, G.: Formal concept analysis in knowledge
processing: a survey on models and techniques. Expert Syst. Appl. 40, 6601–6623 (2013)
85. Pollandt, S.: Fuzzy Begriffe. Springer, Berlin (1997)
86. Popescu, A.: A general approach to fuzzy concepts. Math. Logic Quart. 50(3), 265–280 (2004)
87. Priss, U.: Using FCA to Analyse How Students Learn to Program. In: Cellier, P., Distel, F.,
Ganter, B. (eds.) ICFCA 2013. LNCS, vol. 7880, pp. 216–227. Springer, Heidelberg (2013)
88. Singh, P.K., Kumar, ChA: Bipolar fuzzy graph representation of concept lattice. Inf. Sci. 288,
437–448 (2014)
89. Singh, P.K., Kumar, ChA: A note on bipolar fuzzy graph representation of concept lattice. Int.
J. Comp. Sci. Math. 5(4), 381–393 (2014)
90. Skřivánek, V., Frič, R.: Generalized random events. Int. J. Theor. Phys. 54, 4386–4396 (2015)
91. Snášel, V., Duráková, D., Krajči, S., Vojtáš, P.: Merging concept lattices of α-cuts of fuzzy
contexts. Contrib. Gen. Algebra 14, 155–166 (2004)
92. Valverde-Albacete, F.J., Peláez-Moreno, C.: Towards a generalisation of formal concept anal-
ysis for data mining purposes. In: Missaoui, R., Schmid, J. (eds.) ICFCA 2006. LNCS, vol.
3874, pp. 161–176. Springer, Heidelberg (2006)
On Fuzzy Generalizations of Concept Lattices 103
93. Valverde-Albacete, F.J., Peláez-Moreno, C.: Extending conceptualisation modes for gener-
alised formal concept analysis. Inf. Sci. 181, 1888–1909 (2011)
94. Wille, R.: Restructuring lattice theory: an approach based on hierarchies of concepts. In: Rival,
I. (ed.) Ordered Sets, pp. 445–470. Reidel, Dodrecht-Boston (1982)
Generating Fuzzy Attribute Rules Via
Fuzzy Formal Concept Analysis
1 Introduction
The hidden knowledge databases store is a valuable asset in a wide variety of areas
like stock market prediction [16, 22], disease diagnosis [20, 21] or census data
analysis [7, 18], among others. One of the main techniques in order to represent
the knowledge extracted from a database is by rules, summarizing completely the
information stored in the database. These rules are usually extracted via APRIORI
Partially supported by the State Research Agency (AEI) and the European Regional Develop-
ment Fund (ERDF) projects TIN2015-65845-C3-3-R and TIN2016-76653-P.
algorithms [1], which explore frequent itemsets to select the most frequent and con-
fident rules of the database.
Formal Concept Analysis (FCA) [8] is a mathematical technique which helps
to discover relationships between sets of attributes and objects inside a database,
known as concepts. FCA retrieves the main concepts that a database has, and that
can be useful to obtain a rule set. In the classical FCA [1], only boolean attributes
are considered, leading to a set of rules which consider attributes fully true or false.
A most accurate framework is given by fuzzy generalizations where uncertainty and
noise are present.
The computation of attribute implications in one of the most important research
topics in FCA [4, 5, 10, 11, 15, 28]. The main goal of this paper is to define, in
any fuzzy concept lattice framework, a new kind of fuzzy attribute implications (that
is a rule with attributes that might not be fully true or false) and a base of them,
as a minimum set of rules needed to summarize the whole information present in a
database. A particular case in which intensions are considered in the rules is studied
and an application to clustering the concepts of the concept lattice, which provides
a size reduction mechanism of the concept lattice, is presented.
This paper is structured as follows. In Sect. 2 the preliminary definitions are pre-
sented. Then, Sect. 3 presents fuzzy sc-attribute rules, properties and the notion of
base. Next, Sect. 4 studies the particular case in which intensions are considered in
the rules and Sect. 5 applies these rules in order to provide a clustering in the concept
lattice. Finally, conclusions and future work are shown.
2 Preliminaries
This section presents the required preliminary definitions. In all the definitions a
complete lattice (L , ) and a finite universe U are considered.
The following definition introduces the notion of spanning tree, which will be
used in Sect. 4.
(a set of attributes A, a set of objects B and a fuzzy relation R between them) and a
Galois connection [3, 14, 19] or a family of Galois connections [2, 23].
The theory and results proposed in this paper are applied to all of them. In order
to recall a particular fuzzy concept lattice, mainly for the examples, we will present
the notion of context and the definitions of the concept-forming operators introduced
in this framework.
The first definition introduces the basic operators considered in the concept-
forming operators.
Definition 3 Let (P1 , ≤1 ), (P2 , ≤2 ), (P3 , ≤3 ) be posets and & : P1 × P2 → P3 ,
: P3 × P2 → P1 , : P3 × P1 → P2 be mappings, then(&, , ) is an adjoint
triple with respect to P1 , P2 , P3 if:
where x ∈ P1 , y ∈ P2 and z ∈ P3 .
The posets (P1 , ≤1 ) and (P2 , ≤2 ) considered in the previous definition are com-
plete lattices in the multi-adjoint concept lattice framework [19] and they are denoted
as (L 1 , 1 ) and (L 2 , 2 ). Now the notion of multi-adjoint frame is presented.
Definition 4 A multi-adjoint frame L is a tuple
(L 1 , L 2 , P, 1 , 2 , ≤, &1 , 1 , 1 , . . . , &n , ,
n
n)
M = {g, f | g ∈ L 2B , f ∈ L 1A and g ↑ = f, f ↓ = g}
Example 1 Let (L , , &G ) be a multi-adjoint frame, where &G is the Gödel con-
junctor with respect to L = {0, 0.5, 1}. The context (A, B, R, σ ) is formed by the
sets A = {a1 , a2 , a3 } and B = {b1 , b2 }, the relation R is defined from Table 1 and the
constant mapping σ . In the fuzzy set notation, for each a ∈ A, the expression a/1
will simply be written as a and a/0 will be omitted.
The concept lattice related to the given context is shown in Fig. 1. We can see we
have seven concepts, and the Hasse diagram shows the hierarchy among them.
This section presents a special kind of fuzzy attribute rule and several properties.
From now on, a frame and a context (A, B, R, σ ) will be fixed from which a Galois
connection (↑ , ↓ ) or a family of Galois connections (as in [2, 6, 23]) is defined.
Specifically, we will consider a Galois connection instead of a family, in order to
simplify the notation.
First of all, the classical definitions of support and confidence are provided.
Definition 7 ([1]) The support is defined as the ratio of objects which are related to
a given subset of attributes.
These measures on rules will be extended to the fuzzy one next. The first notion
we will consider is the definition of support, which is a key measure of a fuzzy rule.
Definition 9 The support of f ∈ L 1A in (A, B, R, σ ) is defined as
card(f ↓ )
supp(f ) = (1)
|B|
card({a1 /0.5, a3 }↓ )
supp({a1 /0.5, a3 }) =
|B|
card(b1 ) 1
= =
2 2
The support satisfies interesting properties that are shown next. Since the proof
can straightforwardly be obtained, it is not included.
Once the notion of support has been introduced, the following definition provides
the definition of fuzzy sc-attribute rule and a specific truth value associated with it.
Definition 10 Given two fuzzy subsets of attributes f 1 , f 2 ∈ L 1A , the fuzzy attribute
rule over A from f 1 to f 2 is given by the expression f 2 ←(s,c) f 1 , where s = supp(f1 )
and the confidence c is defined by
supp(f1 ∪ f 2 )
c= (2)
supp(f1 )
If the confidence is 1, the fuzzy sc-attribute rule is called fuzzy sc-attribute implica-
tion, which is denoted by f 2 ⇐ f 1 .
Now, an example of the application of the notion of confidence will be shown.
Example 3 From Example 1, the fuzzy attribute rule {a1 } ←(s,c) {a1 /0.5} can be
considered, where the confidence is computed as:
This mapping satisfies several interesting properties as the following one, which
trivially holds.
The following section will present different properties of the introduced rules.
Given two ordering related fuzzy subsets of attributes, a trivial fuzzy attribute rule
always arises.
Proposition 3 Let f 1 , f 2 ∈ L 1A , where f 1 ≺ f 2 , and the rule f 1 ←(s,c) f 2 , we have
that c = 1.
supp(f1 ∪ f 2 ) supp( f 2 )
Proof The confidence of the rule can be expressed as supp(f2 )
= supp(f2 )
= 1.
supp(f1 ∪ f 2 ) supp( f 2 )
c12 = =
supp(f1 ) supp(f1 )
supp(f1 ∪ f 3 ) supp( f 3 )
c13 = =
supp(f1 ) supp(f1 )
Assuming that f 2 f 3 , we also know that supp(f3 ) ≤ supp( f 2 ). Applying this fact,
the following inequalities are equivalents:
supp(f3 ) ≤ supp( f 2 )
supp(f3 ) supp(f2 )
≤
supp( f 1 ) supp( f 1 )
c13 ≤ c12
Example 4 In the context of Example 1, we will see the previous property in the
rules {a1 } ←(s1 ,c1 ) {a1 /0.5} and {a1 , a3 } ←(s2 ,c2 ) {a1 /0.5}. Hence, we compute the
supports and confidences of both rules:
s1 = supp({a1 /0.5}) = 1
supp({a1 }) 0.75
c1 = = = 0.75
supp({a1 /0.5}) 1
s2 = supp({a1 /0.5}) = 1
supp({a1 , a3 }) 0.25
c2 = = = 0.25
supp({a1 /0.5}) 1
As we can see, c2 ≤ c1 . Now, we will consider other two rules that do not satisfy
the hypotheses in Proposition 4 and we will see that the thesis does not hold either.
Given the rules {a1 , a3 } ←(s3 ,c3 ) {a1 } and {a1 , a2 /0.5} ←(s4 ,c4 ) {a1 }, we have:
s3 = supp({a1 }) = 0.75
supp({a1 , a3 }) 0.25
c3 = = = 0.33
supp({a1 }) 0.75
s4 = supp({a1 }) = 0.75
supp({a1 , a2 /0.5}) 0.5
c4 = = = 0.67
supp({a1 /0.5}) 0.75
112 V. Liñeiro-Barea et al.
The following result shows how the confidence can be derived via transitivity,
based on the idea given in [17].
Theorem 1 Let f 1 , f 2 , f 3 ∈ L 1A , where f 1 ≺ f 2 ≺ f 3 , and the rules f 2 ←(s,c) f 1 ,
f 3 ←(s ,c ) f 2 , f 3 ←(s,c ) f 1 , then we have that c · c = c .
Proof Let f 3 ←(s ,c ) f 1 . The support of the new rule, s can be obtained applying
Eq. 1:
s = supp(f1 ) = s
supp(f1 ∪ f 3 ) supp( f 3 )
c = =
supp(f1 ) supp(f1 )
supp(f1 ∪ f 2 ) supp( f 2 )
c= =
supp(f1 ) supp(f1 )
supp(f ∪ f ) supp( f3 )
c =
2 3
=
supp(f2 ) supp(f2 )
Proof Let f 4 ←(s4 ,c4 ) f 2 . The confidence of the new rule c4 arises from Eq. 2:
supp(f2 ∪ f 4 ) supp( f 4 )
c4 = =
supp(f2 ) supp(f2 )
Given a context, the set of fuzzy attribute rules computed may be huge and it also may
include redundant and not interesting rules for the purpose they are being obtained.
In order to fix that, a base of rules should be obtained.
Definition 11 A fuzzy attribute rule f ←(s,c) f is derived from a set of fuzzy
attribute rules M if a succession { f 1 , f 2 , · · · , f n } of different fuzzy subset of
attributes exists, such that f = f 1 , f = f n , and f i+1 ←(si ,ci ) f i ∈ M or f i ←(si ,ci )
f i+1 ∈ M, for all i ∈ {1, . . . , n − 1}.
Example 5 Considering in Example 1 the set of fuzzy attribute rules M = {{a1 }
←(s12 =1,c12 =0.75) {a1 /0.5}, {a1 , a3 } ←(s23 =0.75,c23 =0.3̂) {a1 }}, we can derive a new rule
from it, {a1 , a3 } ←(s,c) {a1 /0.5}, with:
s = s12 = 1
c = c12 · c23 = 0.75 · 0.3̂ = 0.25
This section defines, in a given concept lattice framework, an sc-attribute rule base
extending the classical case [17] and relates this base to the given concept lattice.
First of all, the notion of fuzzy sc-intension rule is introduced.
Definition 13 Given the set of all intents in the context Int(A, B, R, σ ) = { f ∈
L 1A | f = f ↓↑ }, the fuzzy rule f 2 ←(s,c) f 1 , in which the fuzzy subset of attributes
are intents, that is f 1 , f 2 ∈ Int( A, B, R, σ ), and f 1 ≺ f 2 is called fuzzy sc-intension
rule.
Note that Theorems 1 and 2 show how new fuzzy sc-intension rules can be derived
from a base via transitivity. For example, given f 1 , f 2 , f 3 ∈ Int( A, B, R, σ ), satisfy-
ing f 1 ≺ f 2 ≺ f 3 , and the rules f 2 ←(s,c) f 1 and f 3 ←(s ,c ) f 2 , the rule f 3 ←(s,c ) f 1
can straightforwardly be derived considering c = c · c .
This result provides that only the rules between neighbor intents need to be consid-
ered in a base and gives us a mechanism in order to compute a fuzzy sc-intension rule
base, only considering a minimal subset of rules between neighbor intents removing
cycles in the concept lattice, in other words, obtaining a spanning tree of the Hasse
diagram of the concept lattice.
If we need to compute the confidence of a new rule, we just need to identify the
path followed in the spanning tree and then multiply the confidences of the involved
rules, inverting the value in the rules that we consider in the reverse form. This is
shown in the following example.
5 Application to Clustering
The base of fuzzy sc-intension rules has immediate application in a clustering concept
task. Given a concept lattice M associated with a context (A, B, R), we can naturally
see the confidence of a rule f 2 ←(s,c) f 1 as a ratio of similarity between the concepts
C2 = f 2 ↓ , f 2 and C1 = f 1 ↓ , f 1 . If we establish a threshold δ of confidence, we
can clustered the concepts of M obtaining a set of the similar ones. Note that one
concept can be part of more than one cluster. Hence, the obtained clustering can be
a covering [9, 30].
Given a concept C, the set of concepts which has a confidence greater or equal
than δ with respect to C will be called as the cover in M of C and will be denoted
as [C].
The process of clustering can be summarized in the following steps:
116 V. Liñeiro-Barea et al.
(a) If Step 4 is doing in a sequential process, that is, firstly Steps 2 and 3 are applied
to the maximal element C M1 , then to C M2 , and so on, a partition of M is obtained.
The main weakness is that this partition depend on the considered ordering in
the maximal elements.
(b) If Step 4 is doing in a parallel process, that is, firstly Step 2 is applied to
each maximal element C M1 , . . . , C Mn , and the different, not necessary disjoint
classes [C M1 ], . . . , [C Mn ] are computed, and then Step 3 is applied, a covering
is obtained. In this case, a partition is not obtained, in general.
In the following example, the first option of the clustering algorithm is applied to
a particular concept lattice.
Example 7 Following Example 1, we obtain the fuzzy sc-intension rule base TL
shown in Fig. 2, which provides a maximal TL -weighted spanning tree. The con-
fidence value of each rule is indicated in the edges of the diagram. In each node,
together with the extension and intension of the corresponding concept, a number is
written representing the label of the concept.
If we assume δ = 0.66, we obtain the following partition:
– [C1 ] = {1, 2}
– [C3 ] = {3}
– [C4 ] = {4}
– [C5 ] = {5}
– [C6 ] = {6}
– [C7 ] = {7}
In case of δ = 0.5, we obtain the following partition:
– [C1 ] = {1, 2, 3, 4}
– [C5 ] = {5}
– [C6 ] = {6}
– [C7 ] = {7}
Considering now δ = 0.25, the partition is:
Generating Fuzzy Attribute Rules Via Fuzzy Formal Concept Analysis 117
– [C1 ] = {1, 2, 3, 4, 5, 6}
– [C7 ] = {7}
Note that, in this particular case, since a small concept lattice has been considered,
the algorithm with the second option (b) provides the same (partition) covering.
An interesting consequence of the proposed clustering mechanism is that it pro-
vides a reduction procedure based on the confidence of the fuzzy sc-intension rules.
This is possible because the set of classes always contains the minimum concept.
Moreover, we have that the concepts closer to the top concept provides greater confi-
dences. Therefore, the computation of the top part of the concept lattice is interesting
in order to obtain the most representative fuzzy sc-intension rules.
A new kind of attribute implication has been introduced considering the fuzzy notions
of support and confidence, and the properties of these rules have been studied. These
rules allow to consider fuzzy rules that have a confidence value less than one, c < 1,
which is also interesting since some noise in the data can be mitigated and outlier
data are also considered. The particular case of using intensions in the rules provides
attribute implications with confidence less than 1, which can complement the usual
attribute implications with truth degree 1. Moreover, these rules provide a proce-
dure for clustering the concept lattices and, as a consequence, offer a size reduction
mechanism of the concept lattice.
In the following, a set of future work lines will be presented. First of all, the
performance and the potential of the procedure for information retrieval will be done
118 V. Liñeiro-Barea et al.
with an experimental survey. This will be the main line for future work, allowing to
search for optimizations in big data sets.
Extracting information from a database is the main task fuzzy rules are designed
for. Fuzzy rules can be utilized in Machine Learning scenarios such classification [13,
29] or prediction [12, 25]. Another line for future work will be to consider such
scenarios in order to study the potential of our work in them.
References
1. Agrawal, R., Imielinski, T., Swami, A.: Mining association rules between sets of items in large
databases. In: ACM SIGMOD International Conference on Management of Data, pp. 207–216
(1993)
2. Antoni, L., Krajci, S., Kridlo, O., Macek, B., Pisková, L.: On heterogeneous formal contexts.
Fuzzy Sets Syst. 234, 22–33 (2014)
3. Bělohlávek, R.: Lattices of fixed points of fuzzy Galois connections. Math. Logic Q. 47(1),
111–116 (2001)
4. Belohlavek, R., Cordero, P., Enciso, M., Mora, A., Vychodil, V.: Automated prover for attribute
dependencies in data with grades. Int. J. Approx. Reasoning 70, 51–67 (2016)
5. Belohlávek, R., Vychodil, V.: Attribute dependencies for data with grades ii. Int. J. Gen. Syst.
46(1), 66–92 (2017)
6. Butka, P., Pócs, J.: Generalization of one-sided concept lattices. Comput. Inform. 32(2), 355–
370 (2013)
7. Chertov, O., Aleksandrova, M.: Using association rules for searching levers of influence in
census data. Procedia Soc. Behav. Sci. 73(0), 475–478 (2013), In: Proceedings of the 2nd
International Conference on Integrated Information (IC-ININFO 2012), Budapest, Hungary,
30 Aug–3 Sept 2012
8. Ganter, B., Wille, R.: Formal Concept Analysis: Mathematical Foundations, Springer, New
York (2012)
9. Ge, X., Wang, P., Yun, Z.: The rough membership functions on four types of covering-based
rough sets and their applications. Inf. Sci. 390, 1–14 (2017)
10. Glodeanu, C.V.: Knowledge discovery in data sets with graded attributes. Int. J. Gen. Syst.
45(2), 232–249 (2016)
11. Hill, J., Walkington, H., France, D.: Graduate attributes: implications for higher education
practice and policy. J. Geogr. High. Educ. 40(2), 155–163 (2016)
12. Ikram, A., Qamar, U.: Developing an expert system based on association rules and predicate
logic for earthquake prediction. Knowl. Based Syst. 75, 87–103 (2015)
13. Kianmehr, K., Alhajj, R.: Carsvm: a class association rule-based classification framework and
its application to gene expression data. Artif. Intell. Med. 44(1), 7–25 (2008)
14. Krajči, S.: A generalized concept lattice. Logic J. IGPL 13(5), 543–550 (2005)
15. Kuhr, T., Vychodil, V.: Fuzzy logic programming reduced to reasoning with attribute implica-
tions. Fuzzy Sets Syst. 262, 1–20 (2015)
16. Lu, H., Han, J., Feng, L.: Stock movement prediction and n-dimensional inter-transaction
association rules. In: Proceedings of the ACM SIGMOD Workshop on Research Issues in Data
Mining and Knowledge Discovery, p. 12 (1998)
17. Luxenburger, M.: Implications partielles dans un contexte. Mathématiques, Informatique et
Sciences Humaines 29(113), 35–55 (1991)
18. Malerba, D., Lisi, F.A., Sblendorio, F.: Mining spatial association rules in census data: a rela-
tional approach. In: Proceeding of the ECML/PKDD?02 workshop on Mining Official Data,
University Printing House, Helsinki, Citeseer (2002)
Generating Fuzzy Attribute Rules Via Fuzzy Formal Concept Analysis 119
19. Medina, J., Ojeda-Aciego, M., Ruiz-Calviño, J.: Formal concept analysis via multi-adjoint
concept lattices. Fuzzy Sets Syst. 160(2), 130–144 (2009)
20. Nahar, J., Imam, T., Tickle, K.S., Chen, Y.-P.P.: Association rule mining to detect factors which
contribute to heart disease in males and females. Expert Syst. Appl. 40(4), 1086–1093 (2013)
21. Ordonez, C., Santana, C., de Braal, L.: Discovering interesting association rules in medical
data. In: Proceedings of ACM SIGMOD Workshop on Research Issues on Data Mining and
Knowledge Discovery, pp. 78–85 (2000)
22. Paranjape-Voditel, P., Deshpande, U.: An association rule mining based stock market rec-
ommender system. In: 2011 Second International Conference on Emerging Applications of
Information Technology (EAIT), pp. 21–24, Feb 2011
23. Popescu, A.: A general approach to fuzzy concepts. Math. Logic Q. 50(3), 265–280 (2004)
24. Rai, S., Sharma, S.: Determining minimum spanning tree in an undirected weighted graph.
In: 2015 International Conference on Advances in Computer Engineering and Applications
(ICACEA), pp. 637–642, Mar 2015
25. Rudin, C., Letham, B., Salleb-Aouissi, A., Kogan, E., Madigan, D.: Sequential event prediction
with association rules. In: 24th Annual Conference on Learning Theory (COLT 2011), pp. 615–
634, July 2011
26. Vychodil, V.: Computing sets of graded attribute implications with witnessed non-redundancy
(2015). arxiv: CoRRabs/1511.01640
27. Vychodil, V.: Rational fuzzy attribute logic (2015). arxiv: CoRRabs/1502.07326
28. Vychodil, V.: Computing sets of graded attribute implications with witnessed non-redundancy.
Inf. Sci. 351, 90–100 (2016)
29. Wang, W., Wang, Y., Bañares-Alcántara, R., Cui, Z., Coenen, F.: Application of classification
association rule mining for mammalian mesenchymal stem cell differentiation. In: Perner, P.
(ed.) Advances in Data Mining. Applications and Theoretical Aspects, Volume 5633 of Lecture
Notes in Computer Science, pp. 51–61. Springer, Berlin Heidelberg (2009)
30. Yang, B., Hu, B.O.: On some types of fuzzy covering-based rough sets. Fuzzy Sets Syst. 312,
36–65 (2017) (Theme: Fuzzy Rough Sets)
31. Zadeh, L.A.: Fuzzy logic. Computer 21(4), 83–93 (1988)