Difference between revisions of "Vol-3194/paper41"

From BITPlan ceur-ws Wiki
Jump to navigation Jump to search
(edited by wikiedit)
 
(modified through wikirestore by wf)
Line 1: Line 1:
 
+
=Paper=
 
{{Paper
 
{{Paper
|wikidataid=Q117344879
+
|id=Vol-3194/paper41
 +
|storemode=property
 +
|title=Generating Synthetic Discrete Datasets with Machine Learning
 +
|pdfUrl=https://ceur-ws.org/Vol-3194/paper41.pdf
 +
|volume=Vol-3194
 +
|authors=Giuseppe Manco,Ettore Ritacco,Antonino Rullo,Domenico Saccà,Edoardo Serra
 +
|dblpUrl=https://dblp.org/rec/conf/sebd/0001RRSS22
 
}}
 
}}
 +
==Generating Synthetic Discrete Datasets with Machine Learning==
 +
<pdf width="1500px">https://ceur-ws.org/Vol-3194/paper41.pdf</pdf>
 +
<pre>
 +
Generating Synthetic Discrete Datasets with Machine
 +
Learning
 +
(Discussion Paper)
 +
 +
Giuseppe Manco1 , Ettore Ritacco1 , Antonino Rullo2 , Domenico Saccà2 and
 +
Edoardo Serra3
 +
1
 +
  Institute for High Performance Computing and Networking (ICAR) of the Italian National Research Council (CNR), v. P.
 +
Bucci 8/9C, 87036 Rende (CS), Italy
 +
2
 +
  DIMES Department, University of Calabria, Rende, CS, 87036, Italy
 +
3
 +
  Computer Science Department, Boise State University, Boise, ID 83725, USA
 +
 +
 +
                                        Abstract
 +
                                        The real data are not always available/accessible/sufficient or in many cases they are incomplete and
 +
                                        lacking in semantic content necessary to the definition of optimization processes. In this paper we
 +
                                        discuss about the synthetic data generation under two different perspectives. The core common idea is
 +
                                        to analyze a limited set of real data to learn the main patterns that characterize them and exploit this
 +
                                        knowledge to generate brand new data. The first perspective is constraint-based generation and consists
 +
                                        in generating a synthetic dataset satisfying given support constraints on the real frequent patterns. The
 +
                                        second one is based on probabilistic generative modeling and considers the synthetic generation as a
 +
                                        sampling process from a parametric distribution learned on the real data, typically encoded as a neural
 +
                                        network (e.g. Variational Autoencoders, Generative Adversarial Networks).
 +
 +
                                        Keywords
 +
                                        Synthetic dataset, Data generation, Inverse Frequent Itemset Mining, Constraints-based models, Varia-
 +
                                        tional Autoencoder, Generative Adversarial Networks, Generative models
 +
 +
 +
          This paper is an extended abstract of [1].
 +
 +
 +
1. Introduction
 +
Emerging “Big Data” platforms and applications call for the invention of novel data analysis
 +
techniques that are capable to effectively and efficiently handle large amount of data [2]. There
 +
is therefore an increasing need to use real-life datasets for data-driven experiments but the
 +
scarcity of significant datasets is a critical issue for research papers [3]. Synthetic data generation
 +
can help in this, by reproducing the internal mechanisms and dependencies that justify the
 +
occurrence of some specific pieces of information, and hence being able to replicate them
 +
stochastically. In this paper we focus on the problem of generating high-dimensional discrete
 +
data. Generating such data is a challenge because of the combination of both high dimensionality
 +
and discrete components, which may result in a complex structural domain with lots of variety
 +
and irregularities, not necessarily smooth. Throughout the paper, we shall study approaches to
 +
data generation which rely on the idea that each point in the real domain can be mapped into a
 +
 +
SEBD 2022: The 30th Italian Symposium on Advanced Database Systems, June 19-22, 2022, Tirrenia (PI), Italy
 +
                                      © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
 +
    CEUR
 +
    Workshop
 +
    Proceedings
 +
                  http://ceur-ws.org
 +
                  ISSN 1613-0073
 +
                                      CEUR Workshop Proceedings (CEUR-WS.org)
 +
�Table 1
 +
Transactional Dataset end Frequent Itemsets.
 +
 +
                      𝑎  𝑏  𝑐    𝑑      𝜎𝒟
 +
                      1  1  1    0      40    𝑆𝑖    𝑎    𝑏    𝑐  𝑑        𝜎 𝑆𝑖
 +
                      0  1  1    1      40    𝑆1    1    1    0  0        100
 +
                      1  1  0    0      60    𝑆2    0    1    1  0        100
 +
                      0  1  1    0      20    𝑆3    0    0    1  1          50
 +
                      1  0  1    1      10
 +
                                                (𝑏) Frequent Itemsets with threshold 50
 +
                          (𝑎) Dataset 𝒟
 +
 +
 +
 +
suitable latent space and vice versa. These mappings guarantee a consistency with regards to
 +
the original dataset. At the same time, the manifold in the latent space summarizes the main
 +
characteristics of the data, that can hence be injected into the synthesized data in a controlled
 +
way. We first discuss two main approaches, that are the Inverse Frequent Itemset Mining (IFM)
 +
and probabilistic generative modeling (PGM); finally, a comparison of the results obtained with
 +
some of the presented algorithms are provided.
 +
 +
 +
2. Inverse Frequent Itemset Mining-based Generative Models
 +
Let ℐ be a finite domain of 𝑛 elements, also called items. Any subset 𝐼 ⊆ ℐ is an itemset over
 +
ℐ, also called a transaction. Let 𝒰ℐ denote the set of all itemsets on ℐ; then, |𝒰ℐ | = 2𝑛 . A
 +
(transactional) database 𝒟 is a set of tuples [𝑘, 𝐼], where 𝑘 is the key and 𝐼 is an itemset. The size
 +
|𝒟| of 𝒟 is the total number of its itemsets, i.e., transactions. A transactional database 𝒟 is very
 +
often represented as a bag of itemsets, i.e, the keys are omitted so that tuples are simply itemsets
 +
and may therefore occur duplicated – in this case 𝒟 is also called a transactional dataset. In
 +
the paper we shall also represent an itemset 𝐼 ∈ 𝒟 by its one-hot encoding x , that is a binary
 +
vector of size 𝑛 (the number of items in ℐ) such that its 𝑖-th position 𝑥𝑖 = 1 if the 𝑖-th item in ℐ
 +
is in 𝐼, 0 otherwise. Consequently, 𝒳 = {x 1 , . . . , x 𝜂 }, where 𝜂 = |𝒟|, is the one-hot encoding
 +
of the whole dataset 𝒟. For each itemset 𝐼 ∈ 𝒟, there exist two important measures: (i) the
 +
number of duplicates of 𝐼, denoted as 𝛿 𝒟 (𝐼), that is the number of occurrences of 𝐼 in 𝒟, and (ii)
 +
the support of 𝐼, denoted as 𝜎 𝒟 (𝐼), that
 +
                                        ∑︀ is the sum of the number of duplicates of each itemset
 +
𝐽 in 𝒟 containing 𝐼, i.e., 𝜎 𝒟 (𝐼) = 𝐽∈𝒟∧𝐼⊆𝐽 𝛿 𝒟 (𝐽) . A dataset 𝒟 can be represented in a
 +
succinct format as a set of pairs (𝐼, 𝜎 𝒟 (𝐼)). Given ℐ = {𝑎, 𝑏, 𝑐, 𝑑}, an example of dataset in the
 +
succinct, one-hot format is shown in Table 1-𝑎. We say that 𝐼 is a frequent (resp., infrequent)
 +
itemset in 𝒟 if its support is greater than or equal to (resp., less than) a given threshold. A
 +
classical data mining task over transaction datasets is to detect the set of the frequent/infrequent
 +
itemsets, and a rich literature deals with this topic [4, 5, 6] Given the threshold 50, the frequent
 +
itemsets for the dataset of Table 1-𝑎 are listed in Table 1-𝑏. The perspective of the frequent
 +
itemset mining problem has been later inverted as follows: given a set of itemsets together with
 +
their frequency constraints the goal is to compute, if any, a transaction dataset satisfying the
 +
above constraints. This new problem is called the inverse frequent itemset mining problem (IFM)
 +
algorithms [7], and has been later investigated also in privacy preserving contexts [8]. Given
 +
a set ℐ of items, the IFM problem consists in finding a dataset 𝒟 that satisfies given support
 +
�Table 2
 +
Examples of Feasible Datasets with size 170 for an IFM instance.
 +
 +
            𝑎    𝑏  𝑐  𝑑  𝜎𝒟          𝑎    𝑏  𝑐  𝑑  𝜎𝒟          𝑎    𝑏  𝑐  𝑑    𝜎𝒟
 +
            1    1  1  0  70          1    1  1  0  40          1    1  1  1    30
 +
            0    1  1  1  10          0    1  1  1  40          1    1  1  0    30
 +
            1    1  0  0  30          1    1  0  0  60          0    1  1  1    20
 +
            0    1  1  0  20          0    1  1  0  20          1    1  0  0    40
 +
            0    0  1  1  40          0    0  1  1  10          0    1  1  0    50
 +
                (𝑎) Dataset 𝒟1              (𝑏) Dataset 𝒟2              (𝑐) Dataset 𝒟3
 +
 +
 +
constraints on some itemsets 𝑆𝑖 on ℐ — the set of such itemsets is denoted by 𝑆. The support
 +
constraints are represented as follows: ∀𝑆𝑖 ∈ 𝑆 : 𝜎𝑚𝑖𝑛    𝑖    ≤ 𝜎 𝒟 (𝑆𝑖 ) ≤ 𝜎𝑚𝑎𝑥𝑖  , where 𝜎 𝒟 (𝑆𝑖 ) is
 +
the sum of all number of duplicates of itemsets in 𝒟 containing 𝐼. As an example, consider
 +
ℐ = {𝑎, 𝑏, 𝑐, 𝑑}, 𝑆 = {{𝑎, 𝑏}, {𝑏, 𝑐}, {𝑐, 𝑑}} and the support constraints represented in Table
 +
1-𝑏 – in this example minimal and maximal supports coincide. The itemsets 𝑆1 = {𝑎, 𝑏} and
 +
𝑆2 = {𝑏, 𝑐} must occur in exactly 100 transactions (possibly as their sub-transactions) whereas
 +
the itemset 𝑆3 = {𝑐, 𝑑} must occur in exactly 50 transactions. It is also required that the dataset
 +
size (i.e., the total number of transactions) be 170. The dataset 𝐷1 shown in Table 2-𝑎 is feasible
 +
as it satisfies all constraints: 𝑆1 is satisfied by the transactions {𝑎, 𝑏, 𝑐} and {𝑎, 𝑏}, 𝑆2 by the
 +
transactions {𝑎, 𝑏, 𝑐}, {𝑏, 𝑐, 𝑑} and {𝑏, 𝑐} and 𝑆3 by the transactions {𝑏, 𝑐, 𝑑} and {𝑐, 𝑑}.
 +
  Let 𝑆 ′ be the set of all itemsets that are neither in 𝑆 nor subsets of some itemset in 𝑆. In
 +
the example, 𝑆 ′ consists of {𝑎, 𝑏, 𝑐, 𝑑}, {𝑎, 𝑏, 𝑐}, {𝑎, 𝑏, 𝑑}, {𝑎, 𝑐, 𝑑}, {𝑏, 𝑐, 𝑑}, {𝑎, 𝑐}, {𝑎, 𝑑} and
 +
{𝑏, 𝑑}. IFM does not enforce any constraint on the itemsets in 𝑆 ′ and, therefore, it may happen
 +
that 𝒟 contains additional (and, in some cases, unsuspected or even undesired) frequent itemsets.
 +
In the dataset 𝒟1 of Table 2-𝑎, the itemset {𝑎, 𝑏, 𝑐} is in 𝑆 ′ but it turns out to be frequent with a
 +
support of 70. To remove the anomaly, Guzzo et al. [9] have proposed an alternative formulation,
 +
called IFM𝑆 , that requires that only itemsets in 𝑆 can be included as transactions in 𝒟 and,
 +
therefore, no unexpected frequent itemsets may eventually occur. Obviously, the decision
 +
complexity of this problem is lower as it is NP-complete. Despite the complexity improvement,
 +
the IFM𝑆 formulation has a severe drawback: it is too restrictive in excluding any transaction
 +
besides the ones in 𝑆 as confirmed by the fact that no feasible dataset exists for our running
 +
example. To weaken the tight restrictions of IFM𝑆 , Guzzo et al. [10] proposed a new formulation
 +
of the problem, called IFM with infrequency support constraints (IFMI for short), which admits
 +
transactions in 𝑆 ′ to be in a feasible dataset if their supports are below a given threshold 𝜎 ′ . By
 +
the anti-monotonicity property, the number of infrequency support constraints can be reduced
 +
by applying them only to a subset of 𝑆 ′ consisting of its minimal (inclusion-wise) elements. This
 +
subset, denoted by 𝐵𝑆 ′ , is called the negative border and coincides with the set of all minimal
 +
transversals of the hypergraph 𝐸    ¯ = {ℐ ∖ 𝐼 : 𝐼 ∈ 𝑆} (as defined in [11]). In the example,
 +
𝐵𝑆 ′ = {{𝑎, 𝑐}, {𝑎, 𝑑}, {𝑏, 𝑑}} and the dataset 𝒟2 in Table 2-𝑏 is a feasible dataset for IFMI for
 +
𝜎 ′ = 40. In fact, all infrequency support constraints on 𝐵𝑆 ′ are satisfied as the supports of
 +
{𝑎, 𝑐}, {𝑎, 𝑑}, {𝑏, 𝑑} are respectively 40, 0 and 40. Another possibility to enforce infrequency
 +
constraints is to fix a duplicate threshold 𝛿 ′ so that an itemset in 𝑆 ′ is admitted as transaction
 +
�in a feasible dataset if its number of occurrences is at most 𝛿 ′ . This formulation has been given
 +
in [12] with the name of IFM with infrequency duplicate constraints (IFMD for short). Observe
 +
that duplicate constraints are less restrictive than infrequency constraints in the sense that
 +
some itemset 𝐼 in 𝐵𝑆 ′ may happen to be eventually frequent as it may inherit the supports of
 +
several itemsets in 𝑆 ′ with duplicates below the threshold. For instance, given the threshold
 +
𝛿 ′ = 30, the dataset 𝒟3 in Table 2-𝑐 is a feasible dataset for IFMD . However, the supports of
 +
{𝑎, 𝑐}, {𝑎, 𝑑}, {𝑏, 𝑑} are respectively 60, 30 and 50, thus {𝑎, 𝑐} and {𝑏, 𝑑} are frequent.
 +
 +
 +
3. Machine Learning-based Generative Models
 +
The probabilistic approach to model transactional data (PGM - probabilistic generative modeling)
 +
assumes that in a database 𝒟 the itemsets are modeled as stochastic events: that is, they are
 +
sampled from an unknown true distribution P𝑟 . The analysis of the statistical distribution of the
 +
stochastic events provides insights on the mathematical rules governing the generation process.
 +
The problem hence becomes how to obtain a smooth and reliable estimate of P𝑟 .
 +
  In general, it is convenient to use a parametric model to estimate P𝑟 when the constraints on
 +
the shape of the distribution are known. By associating each observation x with a probability
 +
measure 𝑃 (x |𝜃) ≡ 𝑃𝜃 (x ) where 𝜃 is the set of the distribution parameters, our problem
 +
hence becomes to devise the optimal parameter set 𝜃 that guarantees a reliable approximation
 +
P𝜃 ≈ P𝑟 that can emulate the sampling process x ∼ P𝑟 in a tractable and reliable way. A
 +
clear advantage of a parametric approaches to data generation lies in the insights that it can
 +
provide within the data generation process. They allow to detect the factors governing the
 +
data, providing a meaningful explanation of complex phenomena. In this work, we are focusing
 +
on two state-of-the-art probabilistic approaches: Variational Autoencoders and Generative
 +
Adversarial Networks.
 +
  A Variational Autoencoder (VAE) is an artificial neural architecture that combines tra-
 +
ditional autoencoder architectures [13] with the concept of latent variable modeling [14].
 +
Essentially, we can assume the existence of a a 𝐾-dimensional latent space 𝒵, that can be the
 +
generation engine of the samples in 𝒳 . The transactions 𝒳 = {x 1 , . . . , x 𝜂 } can be modeled
 +
through a chain dependency: (i) given a distribution 𝑃𝜃 (over the parameter set 𝜃) we can
 +
sample z ∈ 𝒵, and (ii) given a z and another distribution 𝑃𝜑 (over the parameter set 𝜑) we
 +
can sample x . According to the maximum likelihood principle, optimal values of 𝜃 and 𝜑 can
 +
be found by trying to maximizing 𝑃 (𝒳 ), but, unfortunately, this optimization is typically an
 +
intractable problem that requires the exploitation of heuristics.
 +
  Variational inference [15] introduces a proposal normal distribution 𝑄𝜆 (z |x ) (over the
 +
parameter set 𝜆), whose purpose is to approximate the true posterior 𝑃 (z |x ). Hence, a VAE
 +
is devised by concatenating two neural networks: an “Encoder”, that maps an input x into
 +
a latent variable z , exploiting 𝑄𝜆 (z |x ), and a “Decoder” that reconstructs x by applying
 +
𝑃𝜑 (x |z ) to z . The gain function, to be optimized to learn the network parameters 𝜆 and
 +
𝜑, is obtained by marginalizing the log-likelihood of 𝑃 (𝒳 ) over z and applying Jensen’s
 +
inequality: log 𝑃 (x ) ≥ Ez ∼𝑄 [log 𝑃 (x |z )] − KL [𝑄(z |x )‖𝑃 (z )], where KL is the Kullback-
 +
Leibler divergence.
 +
  The Decoder can be opportunely exploit to solve the discrete data generation problem. In fact,
 +
�                Embedding plot                              Embedding plot                              Embedding plot
 +
4                                            4                                            4
 +
3                                            3                                            3
 +
2                                            2                                            2
 +
1                                            1                                            1
 +
0                                            0                                            0
 +
1                                            1                                            1
 +
2                                            2                                            2
 +
3                                            3                                            3
 +
4                                            4                                            4
 +
    5  4  3  2      1    0    1  2  3      5  4  3  2      1    0    1  2  3      5  4  3  2      1    0    1  2  3
 +
 +
    (a) Real data mapping in 𝒵              (b) Synthetic data generation                (c) Synthetic data generation
 +
                                                      using MVAE                            using Clustering-variant of
 +
                                                                                                      MVAE
 +
 +
Figure 1: Illustration of the data generation process within VAE.
 +
 +
 +
we can sample as many z as we want, feed them to the Decoder and obtain brand new synthetic
 +
data that is similar to the real one, if the VAE was properly trained. A simple generation approach,
 +
called Multinomial VAE (MVAE), was proposed in [16], where x ∼ Multinomial (𝜋(z )), with
 +
𝜋(z ) = softmax {exp [𝑃𝜑 (z )]} and z ∼ 𝒩 (0 , I 𝐾 ). A more sophisticated approach, that can
 +
better observe the latent space of z and overcome the strong bias of considering only one
 +
standard normal distribution, is implemented by clustering all the z ∼ 𝑄𝜆 (𝒳 ). The generation
 +
of a synthetic x starts by choosing a cluster 𝑐, by exploiting a multinomial sampling according
 +
to the clusters’ densities. Then, we can apply the MVAE sampling to pick up z ∼ 𝒩 (𝜇𝑐 , 𝜎 𝑐 ),
 +
where 𝜇𝑐 and 𝜎 𝑐 are the mean and standard deviation within 𝑐. A comparison of these two
 +
approaches is shown in Figure 1. Unfortunately, the approaches based on maximum likelihood
 +
have been shown to suffer from over generalization [17], especially when data from P𝑟 is limited.
 +
  Generative Adversarial Networks (GANs) [18] propose an alternative modeling which
 +
departs from the maximum likelihood and instead focuses on an alternative optimization strategy.
 +
In order to optimize the weights 𝜃 of a neural network, called Generator G, able to learn the
 +
probability space P𝜃 , Adversarial Networks rely on an auxiliary classifier 𝐷, with weights 𝜑,
 +
trained to discriminate between real and generated data. In practice, optimality can be achieved
 +
when x ∼ P𝜃 is indistinguishable from x ∼ P𝑟 . The training process can be hence devised as a
 +
competitive game, with the generator trying to produce realistic samples, starting from random
 +
samples in the latent space 𝒵, and the classifier focusing on the detection of generated data,
 +
where the objective function is min𝜃 max𝜑 Ex ∼P𝑟 [log 𝐷𝜑 (x )] + Ex ∼P𝜃 [log(1 − 𝐷𝜑 (x ))].
 +
  It can be shown [18] that the adoption of this alternate optimization is equivalent to train 𝜃
 +
to minimize the Jensen-Shannon divergence, that, differently from the approaches based on
 +
maximum likelihood, has the objective of a complete adherence of P𝑟 and P𝜃 . In our context,
 +
transactions are discrete vectors, with x ∈ {0, 1}𝑛 , thus backpropagation does not directly
 +
apply to G and a workaround is needed. The simplest one is to admit a continuous relaxation of
 +
the output of the generator. Just like with the variational autoencoder, the output of 𝐺 can be
 +
modeled a multinomial probability (with no direct sampling channel), rather than a binary vector.
 +
However, there is a major problem with this: the input of the discriminator would be a softmax
 +
distribution from the generated transactions, and a binary vector for the real transactions. As
 +
a result, the discriminator could easily tell them apart, with the result that the GAN would
 +
get stuck in an equilibrium that is not good for the Generator. Different formulations of the
 +
�Table 3
 +
Originary experimental dataset.
 +
                  Tuple ID  𝑖1  𝑖2  𝑖3  𝑖4    nd      itemset  frequent  support
 +
                  𝑡1          1    0    1    1  210    𝑖1            y          640
 +
                  𝑡2          0    1    0    1  140    𝑖2            y          550
 +
                  𝑡3          1    1    1    0  110    𝑖3            y          450
 +
                  𝑡4          1    0    0    0  100    𝑖4            y          500
 +
                  𝑡5          1    1    0    0    90    𝑖1 , 𝑖2      y          270
 +
                  𝑡6          0    0    0    1    80    𝑖1 , 𝑖3      y          380
 +
                  𝑡7          0    1    1    0    70    𝑖1 , 𝑖4      y          280
 +
                  𝑡8          1    1    0    1    70    𝑖3 , 𝑖4      n          210
 +
                  𝑡9          0    1    0    0    70    𝑖2 , 𝑖3      n          180
 +
                  𝑡10        1    0    1    0    60    𝑖2 , 𝑖4      n          210
 +
 +
 +
 +
 +
adversarial training (based on Wasserstein distance [19], namely Wasserstein GANs or WGANs)
 +
can partially mitigate this issue.
 +
 +
 +
4. Comparative Analysis
 +
In this section we compare the two approaches discussed in sections 2 and 3 by analyzing their
 +
behavior on a set of controlled experiments. For these, we use a toy dataset, shown in Table 3,
 +
comprising 4 items upon which 10 patterns are selected. The dataset used in the experiments is
 +
hence built by replicating such patterns with some fixed frequencies. We use the following
 +
approaches to generate the synthetic datasets:
 +
 +
    - IFM: IFM formulation with support ≥ 250.
 +
    - IFM𝐼 : IFM formulation with infrequent itemsets’ support ≤ 210.
 +
    - IFM𝐷 : IFM formulation with transaction duplicates ≤ 100.
 +
    - IFM𝐷𝐼 : IFM𝐷 merged with IFM𝐼 .
 +
    - IFM𝐷𝐼−5% : IFM𝐼 formulation imposing that each transaction has a number of duplicates
 +
      that differ less than 5% from the number of duplicates in the original dataset.
 +
    - VAE: Variational Auto Encoder generator.
 +
    - VAE−{𝑡4 , 𝑡6 }: VAE without the generation of 𝑡4 and 𝑡6 transactions (as visible in Table 3)
 +
      by means of sampling with rejection.
 +
    - IFM𝐷𝐼−5% − {𝑡4 , 𝑡6 }: IFM𝐷𝐼−5% formulation imposing the number of duplicates for 𝑡4
 +
      and 𝑡6 to be zero.
 +
 +
  The purpose of the experiments is to observe the reconstruction process in both methods
 +
and compare the resulting reconstructed datasets. The comparison relies on the transactions
 +
(𝑡1 . . . 𝑡10 in the table) as well as both simple items and item pairs. We evaluate the faithfulness
 +
of the reconstruction in two respects: (1) whether the patterns are reproduced, and (2) whether
 +
their frequencies are faithful. A simple∑︀      metric to measure the reconstruction accuracy is the
 +
                                            |𝐷𝑖 −𝑂𝑖 |
 +
discrepancy 𝒮, computed as 𝒮 =            𝑖∑︀
 +
                                                      , where, for a pattern 𝑖 (either a transaction or an
 +
                                              𝑖 𝑂𝑖
 +
item pair), 𝐷𝑖 and 𝑂𝑖 represent the frequency of the pattern in the reconstructed and original
 +
�Table 4
 +
Discrepancy.
 +
Patterns                IFM    IFM𝐼    IFM𝐷    IFM𝐷𝐼    IFM𝐷𝐼−5%      VAE  VAE−{𝑡4 , 𝑡6 }  IFM𝐷𝐼−5% − {𝑡4 , 𝑡6 }
 +
Transactions        33.00%  12.00%  48.00%  50.00%      4.80%  7.00%        36.00%                  23.10%
 +
Items & item pairs    3.68%    0.82%    1.91%    0.82%      0.11%  3.02%        16.68%                  4.71%
 +
 +
 +
 +
 +
Figure 2: Two-dimensional representation of the latent space
 +
 +
 +
dataset, respectively. Table 4 reports the values of discrepancy, which are further detailed in
 +
Figures 3-5.
 +
  We first analyze the results of the reconstruction for a VAE-based generative model. Figure
 +
3 shows a comparison with IFM𝐷𝐼−5% , the IFM-based formulation which provides the best
 +
performances. The reconstruction provided by IFM𝐷𝐼−5% is extremely faithful, both on the
 +
itemsets and the transactions. This because it enforces that the number of duplicates of each
 +
transaction differs less than 5% from the number of duplicates in the original dataset. Figure 2
 +
shows the details of how the original patterns are mapped into the latent generative space:
 +
the leftmost picture shows the mapping of the original data, and the rightmost shows how
 +
the generation results from a larger region of the latent space. The main advantage of the
 +
approach based on generative modeling through latent variables is that the latter allows to
 +
control the reconstruction process. By acting on the latter we can modify the characteristics of
 +
the reconstructed space. For example, we see that transaction 𝑡11 (the only spurious transaction
 +
generated by VAE) is placed in a specific region, denoted by a red star in the figure. Sampling
 +
repeatedly from that region would allow us to change the overall distribution of the transactions
 +
while still maintaining the itemset distribution. By contrast, the IFM-based approaches are in
 +
general successful in maintaining the itemset distributions. However, they tend to produce a
 +
higher noise with transactions unless not explicitly constrained by the IFM𝐷𝐼−5% formulation
 +
(see Figure 4). This noise can in principle be considered an advantage in specific contexts where
 +
a differentiation from the original dataset is required (e.g., due to privacy concerns).
 +
  In principle, the adoption of IFM allows to implement a reconstruction "by design", by
 +
choosing which itemsets to maintain or suppress. As an evidence, we report the cases of
 +
IFM𝐷𝐼−5% − {𝑡4 , 𝑡6 } and VAE−{𝑡4 , 𝑡6 }, where transaction 𝑡4 and 𝑡6 are removed from the
 +
generation phases. In fact, Figure 5 shows that IFM𝐷𝐼−5% − {𝑡4 , 𝑡6 } keeps the number of
 +
�                            duplicates comparison IFM-VAE                                                                                        support comparison IFM-VAE
 +
          250                                                                                                      700
 +
                                                                                                                  600
 +
          200
 +
                                                                                                                  500
 +
          150                                                                                                      400
 +
          100                                                                                                      300
 +
                                                                                                                  200
 +
          50
 +
                                                                                                                  100
 +
            0                                                                                                        0
 +
                t1    t2  t3        t4  t5    t6    t7        t8        t9    t10 1001 1011 0010                      i1        i2        i3        i4      i1,i2      i1,i3      i1,i4  i3,i4    i2,i3    i2,i4
 +
 +
                                      original        VAE    IFM_DI 5%                                                                              original          VAE      IFM_DI 5%
 +
 +
 +
Figure 3: Details of the discrepancies on the number of duplicates (left) and the itemsets support (right)
 +
between VAE and IFM-based techniques.
 +
 +
                                                                                        duplicates IFM-based techniques
 +
          250
 +
          200
 +
          150
 +
          100
 +
            50
 +
            0
 +
                      t1        t2              t3          t4                  t5          t6          t7          t8                t9          t10            1001          0111        1011          0010
 +
                                                                        original        IFM        IFM_I        IFM_D          IFM_DI          IFM_DI 5%
 +
 +
                                                                                          support IFM-based techniques
 +
                      700
 +
 +
                      600
 +
 +
                      500
 +
 +
                      400
 +
 +
                      300
 +
 +
                      200
 +
 +
                      100
 +
 +
                      0
 +
                                  i1              i2                    i3              i4            i1,i2            i1,i3              i1,i4              i3,i4            i2,i3            i2,i4
 +
                                                                              original    IFM      IFM_I        IFM_D        IFM_DI        IFM_DI 5%
 +
 +
 +
Figure 4: Details of the discrepancies on the number of duplicates (top) and the itemsets support
 +
(bottom) between the original dataset and IFM-based techniques.
 +
 +
 +
duplicates and the supports almost similar to the ones of the original dataset, while VAE−{𝑡4 , 𝑡6 }
 +
changes many of them to remove the two transactions. The approach based on generative
 +
modeling is in general more efficient. However, constraint-based generation is sensitive to the
 +
frequency threshold and a suitable tuning can make these approaches comparable.
 +
  To summarize, these experiments support an underlying intuition: Constraint-based genera-
 +
tion allows more control on the expected outcome at the expense of a higher computational
 +
cost, whereas probabilistic generative models provide more faithful reconstructions but are less
 +
controllable. This essentially means that, without any further modeling artifact (that we do
 +
not consider here), generative models are prone to fail in providing tailored reconstructions
 +
where some patterns can be suppressed and new ones introduced. By contrast, contraint-based
 +
generation is more suitable for reconstructions "by design".
 +
�          duplicates comparison IFM-VAE w/o { t4,t6}                                    support comparison IFM-VAE w/o { t4,t6}
 +
250                                                                          700
 +
200
 +
                                                                            500
 +
150
 +
100                                                                          300
 +
 +
50
 +
                                                                            100
 +
  0
 +
      t1      t2      t3      t5    t7      t8    t9    t10  1001 0011  -100  i1    i2    i3    i4  i1,i2  i1,i3  i1,i4  i3,i4  i2,i3  i2,i4
 +
          original        VAE w/o { t4,t6}        IFM_DI w/o { t4,t6}              original    VAE w/o { t4,t6}      IFM_DI 5% w/o { t4,t6}
 +
 +
 +
 +
Figure 5: Details of the discrepancies on the number of duplicates (left) and the itemsets support (right)
 +
between VAE and IFM-based techniques, without transactions 𝑡4 and 𝑡6 .
 +
 +
 +
5. Conclusions
 +
This paper has provided an overview about state-of-the-art approaches for synthetic transac-
 +
tional data generation. A transaction has been modeled as a high-dimension sparse itemset,
 +
that can be mapped into a binary vector, defined over the item space. The investigated algo-
 +
rithms are (variants of the) Inverse Frequent Itemset Mining (IFM), and Probabilistic Generative
 +
Models (PGMs). According to our analysis, the IFM approaches result to be extremely flexible
 +
and understandable; they enable the control of the data generation procedure by the direct
 +
identification of the discovery patterns to preserve. However, they proved to have extremely
 +
onerous computational costs, making them not feasible in high-dimensional contexts. An
 +
opposite conclusion has been obtained by analyzing PGMs: they are extremely fast and accurate,
 +
but strongly lacking in control, flexibility and understandability. As future work, an interesting
 +
research line is trying to investigate novel methodologies and techniques that are able to take
 +
advantage of IFM and PGMs, by combining their strong points and mitigating their weakness.
 +
Another promising research line is to apply the combination of the two approaches to NoSQL
 +
applications by considering the extension of IFM that has been recently proposed in [12].
 +
 +
 +
References
 +
[1] G. Manco, E. Ritacco, A. Rullo, D. Saccà, E. Serra, Machine learning methods for generating
 +
    high dimensional discrete datasets, WIREs Data Mining Knowl. Discov. 12 (2022). URL:
 +
    https://doi.org/10.1002/widm.1450. doi:10.1002/widm.1450.
 +
[2] C. P. Chen, C.-Y. Zhang, Data-intensive applications, challenges, techniques and tech-
 +
    nologies: A survey on big data, Information Sciences 275 (2014) 314 – 347. doi:https:
 +
    //doi.org/10.1016/j.ins.2014.01.015.
 +
[3] G. Weikum, Where’s the data in the big data wave?, 2013. ACM Sigmod Blog:
 +
    http://wp.sigmod.org/?p=786.
 +
[4] R. Agrawal, T. Imieliński, A. Swami, Mining association rules between sets of items in
 +
    large databases, in: Proceedings of the 1993 ACM SIGMOD International Conference on
 +
    Management of Data, SIGMOD ’93, ACM, New York, NY, USA, 1993, pp. 207–216.
 +
� [5] J. Han, H. Cheng, D. Xin, X. Yan, Frequent pattern mining: current status and fu-
 +
    ture directions, Data Mining and Knowledge Discovery 15 (2007) 55–86. doi:10.1007/
 +
    s10618-006-0059-1.
 +
[6] L. Cagliero, P. Garza, Itemset generalization with cardinality-based constraints, Informa-
 +
    tion Sciences 244 (2013) 161 – 174. doi:https://doi.org/10.1016/j.ins.2013.05.
 +
    008.
 +
[7] T. Mielikainen, On inverse frequent set mining, in: Proceedings of 2nd Workshop on
 +
    Privacy Preserving Data Mining, PPDM ’03, IEEE Computer Society, Washington, DC,
 +
    USA, 2003, pp. 18–23.
 +
[8] R. Agrawal, R. Srikant, Privacy-preserving data mining, in: Proceedings of the 2000 ACM
 +
    SIGMOD International Conference on Management of Data, SIGMOD ’00, ACM, New
 +
    York, NY, USA, 2000, pp. 439–450. doi:10.1145/342009.335438.
 +
[9] A. Guzzo, D. Saccà, E. Serra, An effective approach to inverse frequent set mining, in:
 +
    Proceedings of the 2009 Ninth IEEE International Conference on Data Mining, ICDM ’09,
 +
    IEEE Computer Society, Washington, DC, USA, 2009, pp. 806–811. doi:10.1109/ICDM.
 +
    2009.123.
 +
[10] A. Guzzo, L. Moccia, D. Saccà, E. Serra, Solving inverse frequent itemset mining with
 +
    infrequency constraints via large-scale linear programs, ACM Transactions on Knowledge
 +
    Discovery from Data 7 (2013) 18:1–18:39. doi:10.1145/2541268.2541271.
 +
[11] D. Gunopulos, R. Khardon, H. Mannila, H. Toivonen, Data mining, hypergraph transversals,
 +
    and machine learning, in: Proceedings of the 16-th ACM SIGMOD-SIGACT-SIGART
 +
    Symposium on Principles of Database Systems, PODS ’97, 1997, pp. 209–216. doi:10.
 +
    1145/263661.263684.
 +
[12] D. Saccà, E. Serra, A. Rullo, Extending inverse frequent itemsets mining to generate realistic
 +
    datasets: complexity, accuracy and emerging applications, Data Mining and Knowledge
 +
    Discovery 33 (2019) 1736–1774. doi:10.1007/s10618-019-00643-1.
 +
[13] P. Baldi, Autoencoders, unsupervised learning, and deep architectures, in: In Proceedings
 +
    of the International Conference on Unsupervised and Transfer Learning workshop (UTLW),
 +
    volume 27, PMLR, 2011, pp. 37–49.
 +
[14] K. P. Murphy, Machine Learning: A Probabilistic Perspective, The MIT Press, 2012.
 +
[15] D. M. Blei, A. Kucukelbir, J. D. McAuliffe, Variational inference: A review for statisti-
 +
    cians, Journal of the American Statistical Association 112 (2017) 859–877. doi:10.1080/
 +
    01621459.2017.1285773.
 +
[16] D. Liang, R. G. Krishnan, M. Hoffman, T. Jebara, Variational autoencoders for collaborative
 +
    filtering, in: Proceedings of the 2018 World Wide Web Conference, WWW ’18, 2018, pp.
 +
    689–698.
 +
[17] L. Theis, A. van den Oord, M. Bethge, A note on the evaluation of generative models,
 +
    in: International Conference on Learning Representations, 2016. doi:10.48550/arXiv.
 +
    1511.01844.
 +
[18] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville,
 +
    Y. Bengio, Generative adversarial nets, in: Advances in Neural Information Processing
 +
    Systems, volume 27, 2014.
 +
[19] M. Arjovsky, S. Chintala, L. Bottou, Wasserstein generative adversarial networks, in:
 +
    Proceedings of the 34th International Conference on Machine Learning, 2017, pp. 214–223.
 +
 +
</pre>

Revision as of 18:51, 30 March 2023

Paper

Paper
edit
description  
id  Vol-3194/paper41
wikidataid  →Q117344879
title  Generating Synthetic Discrete Datasets with Machine Learning
pdfUrl  https://ceur-ws.org/Vol-3194/paper41.pdf
dblpUrl  https://dblp.org/rec/conf/sebd/0001RRSS22
volume  Vol-3194→Vol-3194
session  →

Generating Synthetic Discrete Datasets with Machine Learning

load PDF

Generating Synthetic Discrete Datasets with Machine
Learning
(Discussion Paper)

Giuseppe Manco1 , Ettore Ritacco1 , Antonino Rullo2 , Domenico Saccà2 and
Edoardo Serra3
1
  Institute for High Performance Computing and Networking (ICAR) of the Italian National Research Council (CNR), v. P.
Bucci 8/9C, 87036 Rende (CS), Italy
2
  DIMES Department, University of Calabria, Rende, CS, 87036, Italy
3
  Computer Science Department, Boise State University, Boise, ID 83725, USA


                                         Abstract
                                         The real data are not always available/accessible/sufficient or in many cases they are incomplete and
                                         lacking in semantic content necessary to the definition of optimization processes. In this paper we
                                         discuss about the synthetic data generation under two different perspectives. The core common idea is
                                         to analyze a limited set of real data to learn the main patterns that characterize them and exploit this
                                         knowledge to generate brand new data. The first perspective is constraint-based generation and consists
                                         in generating a synthetic dataset satisfying given support constraints on the real frequent patterns. The
                                         second one is based on probabilistic generative modeling and considers the synthetic generation as a
                                         sampling process from a parametric distribution learned on the real data, typically encoded as a neural
                                         network (e.g. Variational Autoencoders, Generative Adversarial Networks).

                                         Keywords
                                         Synthetic dataset, Data generation, Inverse Frequent Itemset Mining, Constraints-based models, Varia-
                                         tional Autoencoder, Generative Adversarial Networks, Generative models


           This paper is an extended abstract of [1].


1. Introduction
Emerging “Big Data” platforms and applications call for the invention of novel data analysis
techniques that are capable to effectively and efficiently handle large amount of data [2]. There
is therefore an increasing need to use real-life datasets for data-driven experiments but the
scarcity of significant datasets is a critical issue for research papers [3]. Synthetic data generation
can help in this, by reproducing the internal mechanisms and dependencies that justify the
occurrence of some specific pieces of information, and hence being able to replicate them
stochastically. In this paper we focus on the problem of generating high-dimensional discrete
data. Generating such data is a challenge because of the combination of both high dimensionality
and discrete components, which may result in a complex structural domain with lots of variety
and irregularities, not necessarily smooth. Throughout the paper, we shall study approaches to
data generation which rely on the idea that each point in the real domain can be mapped into a

SEBD 2022: The 30th Italian Symposium on Advanced Database Systems, June 19-22, 2022, Tirrenia (PI), Italy
                                       © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
    CEUR
    Workshop
    Proceedings
                  http://ceur-ws.org
                  ISSN 1613-0073
                                       CEUR Workshop Proceedings (CEUR-WS.org)
�Table 1
Transactional Dataset end Frequent Itemsets.

                      𝑎   𝑏   𝑐    𝑑      𝜎𝒟
                      1   1   1    0      40    𝑆𝑖    𝑎    𝑏    𝑐   𝑑         𝜎 𝑆𝑖
                      0   1   1    1      40    𝑆1    1    1    0   0         100
                      1   1   0    0      60    𝑆2    0    1    1   0         100
                      0   1   1    0      20    𝑆3    0    0    1   1          50
                      1   0   1    1      10
                                                (𝑏) Frequent Itemsets with threshold 50
                          (𝑎) Dataset 𝒟



suitable latent space and vice versa. These mappings guarantee a consistency with regards to
the original dataset. At the same time, the manifold in the latent space summarizes the main
characteristics of the data, that can hence be injected into the synthesized data in a controlled
way. We first discuss two main approaches, that are the Inverse Frequent Itemset Mining (IFM)
and probabilistic generative modeling (PGM); finally, a comparison of the results obtained with
some of the presented algorithms are provided.


2. Inverse Frequent Itemset Mining-based Generative Models
Let ℐ be a finite domain of 𝑛 elements, also called items. Any subset 𝐼 ⊆ ℐ is an itemset over
ℐ, also called a transaction. Let 𝒰ℐ denote the set of all itemsets on ℐ; then, |𝒰ℐ | = 2𝑛 . A
(transactional) database 𝒟 is a set of tuples [𝑘, 𝐼], where 𝑘 is the key and 𝐼 is an itemset. The size
|𝒟| of 𝒟 is the total number of its itemsets, i.e., transactions. A transactional database 𝒟 is very
often represented as a bag of itemsets, i.e, the keys are omitted so that tuples are simply itemsets
and may therefore occur duplicated – in this case 𝒟 is also called a transactional dataset. In
the paper we shall also represent an itemset 𝐼 ∈ 𝒟 by its one-hot encoding x , that is a binary
vector of size 𝑛 (the number of items in ℐ) such that its 𝑖-th position 𝑥𝑖 = 1 if the 𝑖-th item in ℐ
is in 𝐼, 0 otherwise. Consequently, 𝒳 = {x 1 , . . . , x 𝜂 }, where 𝜂 = |𝒟|, is the one-hot encoding
of the whole dataset 𝒟. For each itemset 𝐼 ∈ 𝒟, there exist two important measures: (i) the
number of duplicates of 𝐼, denoted as 𝛿 𝒟 (𝐼), that is the number of occurrences of 𝐼 in 𝒟, and (ii)
the support of 𝐼, denoted as 𝜎 𝒟 (𝐼), that
                                        ∑︀ is the sum of the number of duplicates of each itemset
𝐽 in 𝒟 containing 𝐼, i.e., 𝜎 𝒟 (𝐼) = 𝐽∈𝒟∧𝐼⊆𝐽 𝛿 𝒟 (𝐽) . A dataset 𝒟 can be represented in a
succinct format as a set of pairs (𝐼, 𝜎 𝒟 (𝐼)). Given ℐ = {𝑎, 𝑏, 𝑐, 𝑑}, an example of dataset in the
succinct, one-hot format is shown in Table 1-𝑎. We say that 𝐼 is a frequent (resp., infrequent)
itemset in 𝒟 if its support is greater than or equal to (resp., less than) a given threshold. A
classical data mining task over transaction datasets is to detect the set of the frequent/infrequent
itemsets, and a rich literature deals with this topic [4, 5, 6] Given the threshold 50, the frequent
itemsets for the dataset of Table 1-𝑎 are listed in Table 1-𝑏. The perspective of the frequent
itemset mining problem has been later inverted as follows: given a set of itemsets together with
their frequency constraints the goal is to compute, if any, a transaction dataset satisfying the
above constraints. This new problem is called the inverse frequent itemset mining problem (IFM)
algorithms [7], and has been later investigated also in privacy preserving contexts [8]. Given
a set ℐ of items, the IFM problem consists in finding a dataset 𝒟 that satisfies given support
�Table 2
Examples of Feasible Datasets with size 170 for an IFM instance.

             𝑎    𝑏   𝑐   𝑑   𝜎𝒟          𝑎    𝑏   𝑐   𝑑   𝜎𝒟          𝑎    𝑏   𝑐   𝑑    𝜎𝒟
             1    1   1   0   70          1    1   1   0   40          1    1   1   1    30
             0    1   1   1   10          0    1   1   1   40          1    1   1   0    30
             1    1   0   0   30          1    1   0   0   60          0    1   1   1    20
             0    1   1   0   20          0    1   1   0   20          1    1   0   0    40
             0    0   1   1   40          0    0   1   1   10          0    1   1   0    50
                 (𝑎) Dataset 𝒟1               (𝑏) Dataset 𝒟2               (𝑐) Dataset 𝒟3


constraints on some itemsets 𝑆𝑖 on ℐ — the set of such itemsets is denoted by 𝑆. The support
constraints are represented as follows: ∀𝑆𝑖 ∈ 𝑆 : 𝜎𝑚𝑖𝑛     𝑖    ≤ 𝜎 𝒟 (𝑆𝑖 ) ≤ 𝜎𝑚𝑎𝑥𝑖   , where 𝜎 𝒟 (𝑆𝑖 ) is
the sum of all number of duplicates of itemsets in 𝒟 containing 𝐼. As an example, consider
ℐ = {𝑎, 𝑏, 𝑐, 𝑑}, 𝑆 = {{𝑎, 𝑏}, {𝑏, 𝑐}, {𝑐, 𝑑}} and the support constraints represented in Table
1-𝑏 – in this example minimal and maximal supports coincide. The itemsets 𝑆1 = {𝑎, 𝑏} and
𝑆2 = {𝑏, 𝑐} must occur in exactly 100 transactions (possibly as their sub-transactions) whereas
the itemset 𝑆3 = {𝑐, 𝑑} must occur in exactly 50 transactions. It is also required that the dataset
size (i.e., the total number of transactions) be 170. The dataset 𝐷1 shown in Table 2-𝑎 is feasible
as it satisfies all constraints: 𝑆1 is satisfied by the transactions {𝑎, 𝑏, 𝑐} and {𝑎, 𝑏}, 𝑆2 by the
transactions {𝑎, 𝑏, 𝑐}, {𝑏, 𝑐, 𝑑} and {𝑏, 𝑐} and 𝑆3 by the transactions {𝑏, 𝑐, 𝑑} and {𝑐, 𝑑}.
   Let 𝑆 ′ be the set of all itemsets that are neither in 𝑆 nor subsets of some itemset in 𝑆. In
the example, 𝑆 ′ consists of {𝑎, 𝑏, 𝑐, 𝑑}, {𝑎, 𝑏, 𝑐}, {𝑎, 𝑏, 𝑑}, {𝑎, 𝑐, 𝑑}, {𝑏, 𝑐, 𝑑}, {𝑎, 𝑐}, {𝑎, 𝑑} and
{𝑏, 𝑑}. IFM does not enforce any constraint on the itemsets in 𝑆 ′ and, therefore, it may happen
that 𝒟 contains additional (and, in some cases, unsuspected or even undesired) frequent itemsets.
In the dataset 𝒟1 of Table 2-𝑎, the itemset {𝑎, 𝑏, 𝑐} is in 𝑆 ′ but it turns out to be frequent with a
support of 70. To remove the anomaly, Guzzo et al. [9] have proposed an alternative formulation,
called IFM𝑆 , that requires that only itemsets in 𝑆 can be included as transactions in 𝒟 and,
therefore, no unexpected frequent itemsets may eventually occur. Obviously, the decision
complexity of this problem is lower as it is NP-complete. Despite the complexity improvement,
the IFM𝑆 formulation has a severe drawback: it is too restrictive in excluding any transaction
besides the ones in 𝑆 as confirmed by the fact that no feasible dataset exists for our running
example. To weaken the tight restrictions of IFM𝑆 , Guzzo et al. [10] proposed a new formulation
of the problem, called IFM with infrequency support constraints (IFMI for short), which admits
transactions in 𝑆 ′ to be in a feasible dataset if their supports are below a given threshold 𝜎 ′ . By
the anti-monotonicity property, the number of infrequency support constraints can be reduced
by applying them only to a subset of 𝑆 ′ consisting of its minimal (inclusion-wise) elements. This
subset, denoted by 𝐵𝑆 ′ , is called the negative border and coincides with the set of all minimal
transversals of the hypergraph 𝐸     ¯ = {ℐ ∖ 𝐼 : 𝐼 ∈ 𝑆} (as defined in [11]). In the example,
𝐵𝑆 ′ = {{𝑎, 𝑐}, {𝑎, 𝑑}, {𝑏, 𝑑}} and the dataset 𝒟2 in Table 2-𝑏 is a feasible dataset for IFMI for
𝜎 ′ = 40. In fact, all infrequency support constraints on 𝐵𝑆 ′ are satisfied as the supports of
{𝑎, 𝑐}, {𝑎, 𝑑}, {𝑏, 𝑑} are respectively 40, 0 and 40. Another possibility to enforce infrequency
constraints is to fix a duplicate threshold 𝛿 ′ so that an itemset in 𝑆 ′ is admitted as transaction
�in a feasible dataset if its number of occurrences is at most 𝛿 ′ . This formulation has been given
in [12] with the name of IFM with infrequency duplicate constraints (IFMD for short). Observe
that duplicate constraints are less restrictive than infrequency constraints in the sense that
some itemset 𝐼 in 𝐵𝑆 ′ may happen to be eventually frequent as it may inherit the supports of
several itemsets in 𝑆 ′ with duplicates below the threshold. For instance, given the threshold
𝛿 ′ = 30, the dataset 𝒟3 in Table 2-𝑐 is a feasible dataset for IFMD . However, the supports of
{𝑎, 𝑐}, {𝑎, 𝑑}, {𝑏, 𝑑} are respectively 60, 30 and 50, thus {𝑎, 𝑐} and {𝑏, 𝑑} are frequent.


3. Machine Learning-based Generative Models
The probabilistic approach to model transactional data (PGM - probabilistic generative modeling)
assumes that in a database 𝒟 the itemsets are modeled as stochastic events: that is, they are
sampled from an unknown true distribution P𝑟 . The analysis of the statistical distribution of the
stochastic events provides insights on the mathematical rules governing the generation process.
The problem hence becomes how to obtain a smooth and reliable estimate of P𝑟 .
   In general, it is convenient to use a parametric model to estimate P𝑟 when the constraints on
the shape of the distribution are known. By associating each observation x with a probability
measure 𝑃 (x |𝜃) ≡ 𝑃𝜃 (x ) where 𝜃 is the set of the distribution parameters, our problem
hence becomes to devise the optimal parameter set 𝜃 that guarantees a reliable approximation
P𝜃 ≈ P𝑟 that can emulate the sampling process x ∼ P𝑟 in a tractable and reliable way. A
clear advantage of a parametric approaches to data generation lies in the insights that it can
provide within the data generation process. They allow to detect the factors governing the
data, providing a meaningful explanation of complex phenomena. In this work, we are focusing
on two state-of-the-art probabilistic approaches: Variational Autoencoders and Generative
Adversarial Networks.
   A Variational Autoencoder (VAE) is an artificial neural architecture that combines tra-
ditional autoencoder architectures [13] with the concept of latent variable modeling [14].
Essentially, we can assume the existence of a a 𝐾-dimensional latent space 𝒵, that can be the
generation engine of the samples in 𝒳 . The transactions 𝒳 = {x 1 , . . . , x 𝜂 } can be modeled
through a chain dependency: (i) given a distribution 𝑃𝜃 (over the parameter set 𝜃) we can
sample z ∈ 𝒵, and (ii) given a z and another distribution 𝑃𝜑 (over the parameter set 𝜑) we
can sample x . According to the maximum likelihood principle, optimal values of 𝜃 and 𝜑 can
be found by trying to maximizing 𝑃 (𝒳 ), but, unfortunately, this optimization is typically an
intractable problem that requires the exploitation of heuristics.
   Variational inference [15] introduces a proposal normal distribution 𝑄𝜆 (z |x ) (over the
parameter set 𝜆), whose purpose is to approximate the true posterior 𝑃 (z |x ). Hence, a VAE
is devised by concatenating two neural networks: an “Encoder”, that maps an input x into
a latent variable z , exploiting 𝑄𝜆 (z |x ), and a “Decoder” that reconstructs x by applying
𝑃𝜑 (x |z ) to z . The gain function, to be optimized to learn the network parameters 𝜆 and
𝜑, is obtained by marginalizing the log-likelihood of 𝑃 (𝒳 ) over z and applying Jensen’s
inequality: log 𝑃 (x ) ≥ Ez ∼𝑄 [log 𝑃 (x |z )] − KL [𝑄(z |x )‖𝑃 (z )], where KL is the Kullback-
Leibler divergence.
   The Decoder can be opportunely exploit to solve the discrete data generation problem. In fact,
�                 Embedding plot                               Embedding plot                               Embedding plot
 4                                            4                                            4
 3                                            3                                            3
 2                                            2                                            2
 1                                            1                                            1
 0                                            0                                            0
 1                                            1                                            1
 2                                            2                                            2
 3                                            3                                            3
 4                                            4                                            4
     5   4   3   2      1    0    1   2   3       5   4   3   2      1    0    1   2   3       5   4   3   2      1    0    1   2   3

     (a) Real data mapping in 𝒵               (b) Synthetic data generation                (c) Synthetic data generation
                                                      using MVAE                            using Clustering-variant of
                                                                                                      MVAE

Figure 1: Illustration of the data generation process within VAE.


we can sample as many z as we want, feed them to the Decoder and obtain brand new synthetic
data that is similar to the real one, if the VAE was properly trained. A simple generation approach,
called Multinomial VAE (MVAE), was proposed in [16], where x ∼ Multinomial (𝜋(z )), with
𝜋(z ) = softmax {exp [𝑃𝜑 (z )]} and z ∼ 𝒩 (0 , I 𝐾 ). A more sophisticated approach, that can
better observe the latent space of z and overcome the strong bias of considering only one
standard normal distribution, is implemented by clustering all the z ∼ 𝑄𝜆 (𝒳 ). The generation
of a synthetic x starts by choosing a cluster 𝑐, by exploiting a multinomial sampling according
to the clusters’ densities. Then, we can apply the MVAE sampling to pick up z ∼ 𝒩 (𝜇𝑐 , 𝜎 𝑐 ),
where 𝜇𝑐 and 𝜎 𝑐 are the mean and standard deviation within 𝑐. A comparison of these two
approaches is shown in Figure 1. Unfortunately, the approaches based on maximum likelihood
have been shown to suffer from over generalization [17], especially when data from P𝑟 is limited.
   Generative Adversarial Networks (GANs) [18] propose an alternative modeling which
departs from the maximum likelihood and instead focuses on an alternative optimization strategy.
In order to optimize the weights 𝜃 of a neural network, called Generator G, able to learn the
probability space P𝜃 , Adversarial Networks rely on an auxiliary classifier 𝐷, with weights 𝜑,
trained to discriminate between real and generated data. In practice, optimality can be achieved
when x ∼ P𝜃 is indistinguishable from x ∼ P𝑟 . The training process can be hence devised as a
competitive game, with the generator trying to produce realistic samples, starting from random
samples in the latent space 𝒵, and the classifier focusing on the detection of generated data,
where the objective function is min𝜃 max𝜑 Ex ∼P𝑟 [log 𝐷𝜑 (x )] + Ex ∼P𝜃 [log(1 − 𝐷𝜑 (x ))].
   It can be shown [18] that the adoption of this alternate optimization is equivalent to train 𝜃
to minimize the Jensen-Shannon divergence, that, differently from the approaches based on
maximum likelihood, has the objective of a complete adherence of P𝑟 and P𝜃 . In our context,
transactions are discrete vectors, with x ∈ {0, 1}𝑛 , thus backpropagation does not directly
apply to G and a workaround is needed. The simplest one is to admit a continuous relaxation of
the output of the generator. Just like with the variational autoencoder, the output of 𝐺 can be
modeled a multinomial probability (with no direct sampling channel), rather than a binary vector.
However, there is a major problem with this: the input of the discriminator would be a softmax
distribution from the generated transactions, and a binary vector for the real transactions. As
a result, the discriminator could easily tell them apart, with the result that the GAN would
get stuck in an equilibrium that is not good for the Generator. Different formulations of the
�Table 3
Originary experimental dataset.
                   Tuple ID   𝑖1   𝑖2   𝑖3   𝑖4    nd      itemset   frequent   support
                   𝑡1          1    0    1    1   210     𝑖1            y           640
                   𝑡2          0    1    0    1   140     𝑖2            y           550
                   𝑡3          1    1    1    0   110     𝑖3            y           450
                   𝑡4          1    0    0    0   100     𝑖4            y           500
                   𝑡5          1    1    0    0    90     𝑖1 , 𝑖2       y           270
                   𝑡6          0    0    0    1    80     𝑖1 , 𝑖3       y           380
                   𝑡7          0    1    1    0    70     𝑖1 , 𝑖4       y           280
                   𝑡8          1    1    0    1    70     𝑖3 , 𝑖4       n           210
                   𝑡9          0    1    0    0    70     𝑖2 , 𝑖3       n           180
                   𝑡10         1    0    1    0    60     𝑖2 , 𝑖4       n           210




adversarial training (based on Wasserstein distance [19], namely Wasserstein GANs or WGANs)
can partially mitigate this issue.


4. Comparative Analysis
In this section we compare the two approaches discussed in sections 2 and 3 by analyzing their
behavior on a set of controlled experiments. For these, we use a toy dataset, shown in Table 3,
comprising 4 items upon which 10 patterns are selected. The dataset used in the experiments is
hence built by replicating such patterns with some fixed frequencies. We use the following
approaches to generate the synthetic datasets:

    - IFM: IFM formulation with support ≥ 250.
    - IFM𝐼 : IFM formulation with infrequent itemsets’ support ≤ 210.
    - IFM𝐷 : IFM formulation with transaction duplicates ≤ 100.
    - IFM𝐷𝐼 : IFM𝐷 merged with IFM𝐼 .
    - IFM𝐷𝐼−5% : IFM𝐼 formulation imposing that each transaction has a number of duplicates
      that differ less than 5% from the number of duplicates in the original dataset.
    - VAE: Variational Auto Encoder generator.
    - VAE−{𝑡4 , 𝑡6 }: VAE without the generation of 𝑡4 and 𝑡6 transactions (as visible in Table 3)
      by means of sampling with rejection.
    - IFM𝐷𝐼−5% − {𝑡4 , 𝑡6 }: IFM𝐷𝐼−5% formulation imposing the number of duplicates for 𝑡4
      and 𝑡6 to be zero.

   The purpose of the experiments is to observe the reconstruction process in both methods
and compare the resulting reconstructed datasets. The comparison relies on the transactions
(𝑡1 . . . 𝑡10 in the table) as well as both simple items and item pairs. We evaluate the faithfulness
of the reconstruction in two respects: (1) whether the patterns are reproduced, and (2) whether
their frequencies are faithful. A simple∑︀      metric to measure the reconstruction accuracy is the
                                            |𝐷𝑖 −𝑂𝑖 |
discrepancy 𝒮, computed as 𝒮 =            𝑖∑︀
                                                      , where, for a pattern 𝑖 (either a transaction or an
                                              𝑖 𝑂𝑖
item pair), 𝐷𝑖 and 𝑂𝑖 represent the frequency of the pattern in the reconstructed and original
�Table 4
Discrepancy.
 Patterns                IFM    IFM𝐼     IFM𝐷    IFM𝐷𝐼    IFM𝐷𝐼−5%      VAE   VAE−{𝑡4 , 𝑡6 }   IFM𝐷𝐼−5% − {𝑡4 , 𝑡6 }
 Transactions         33.00%   12.00%   48.00%   50.00%       4.80%   7.00%        36.00%                  23.10%
 Items & item pairs    3.68%    0.82%    1.91%    0.82%       0.11%   3.02%        16.68%                   4.71%




Figure 2: Two-dimensional representation of the latent space


dataset, respectively. Table 4 reports the values of discrepancy, which are further detailed in
Figures 3-5.
   We first analyze the results of the reconstruction for a VAE-based generative model. Figure
3 shows a comparison with IFM𝐷𝐼−5% , the IFM-based formulation which provides the best
performances. The reconstruction provided by IFM𝐷𝐼−5% is extremely faithful, both on the
itemsets and the transactions. This because it enforces that the number of duplicates of each
transaction differs less than 5% from the number of duplicates in the original dataset. Figure 2
shows the details of how the original patterns are mapped into the latent generative space:
the leftmost picture shows the mapping of the original data, and the rightmost shows how
the generation results from a larger region of the latent space. The main advantage of the
approach based on generative modeling through latent variables is that the latter allows to
control the reconstruction process. By acting on the latter we can modify the characteristics of
the reconstructed space. For example, we see that transaction 𝑡11 (the only spurious transaction
generated by VAE) is placed in a specific region, denoted by a red star in the figure. Sampling
repeatedly from that region would allow us to change the overall distribution of the transactions
while still maintaining the itemset distribution. By contrast, the IFM-based approaches are in
general successful in maintaining the itemset distributions. However, they tend to produce a
higher noise with transactions unless not explicitly constrained by the IFM𝐷𝐼−5% formulation
(see Figure 4). This noise can in principle be considered an advantage in specific contexts where
a differentiation from the original dataset is required (e.g., due to privacy concerns).
   In principle, the adoption of IFM allows to implement a reconstruction "by design", by
choosing which itemsets to maintain or suppress. As an evidence, we report the cases of
IFM𝐷𝐼−5% − {𝑡4 , 𝑡6 } and VAE−{𝑡4 , 𝑡6 }, where transaction 𝑡4 and 𝑡6 are removed from the
generation phases. In fact, Figure 5 shows that IFM𝐷𝐼−5% − {𝑡4 , 𝑡6 } keeps the number of
�                             duplicates comparison IFM-VAE                                                                                        support comparison IFM-VAE
          250                                                                                                      700
                                                                                                                   600
          200
                                                                                                                   500
          150                                                                                                      400
          100                                                                                                      300
                                                                                                                   200
           50
                                                                                                                   100
            0                                                                                                        0
                 t1     t2   t3        t4   t5     t6     t7        t8        t9    t10 1001 1011 0010                       i1         i2         i3        i4       i1,i2      i1,i3      i1,i4   i3,i4    i2,i3    i2,i4

                                       original         VAE     IFM_DI 5%                                                                               original          VAE       IFM_DI 5%


Figure 3: Details of the discrepancies on the number of duplicates (left) and the itemsets support (right)
between VAE and IFM-based techniques.

                                                                                         duplicates IFM-based techniques
           250
           200
           150
           100
            50
            0
                       t1         t2              t3           t4                  t5           t6           t7           t8                 t9           t10             1001           0111         1011           0010
                                                                         original         IFM        IFM_I        IFM_D           IFM_DI           IFM_DI 5%

                                                                                           support IFM-based techniques
                      700

                      600

                      500

                      400

                      300

                      200

                      100

                       0
                                  i1               i2                    i3               i4            i1,i2             i1,i3               i1,i4               i3,i4             i2,i3            i2,i4
                                                                              original     IFM       IFM_I        IFM_D        IFM_DI         IFM_DI 5%


Figure 4: Details of the discrepancies on the number of duplicates (top) and the itemsets support
(bottom) between the original dataset and IFM-based techniques.


duplicates and the supports almost similar to the ones of the original dataset, while VAE−{𝑡4 , 𝑡6 }
changes many of them to remove the two transactions. The approach based on generative
modeling is in general more efficient. However, constraint-based generation is sensitive to the
frequency threshold and a suitable tuning can make these approaches comparable.
   To summarize, these experiments support an underlying intuition: Constraint-based genera-
tion allows more control on the expected outcome at the expense of a higher computational
cost, whereas probabilistic generative models provide more faithful reconstructions but are less
controllable. This essentially means that, without any further modeling artifact (that we do
not consider here), generative models are prone to fail in providing tailored reconstructions
where some patterns can be suppressed and new ones introduced. By contrast, contraint-based
generation is more suitable for reconstructions "by design".
�           duplicates comparison IFM-VAE w/o { t4,t6}                                    support comparison IFM-VAE w/o { t4,t6}
250                                                                          700
200
                                                                             500
150
100                                                                          300

 50
                                                                             100
  0
      t1      t2      t3       t5    t7       t8     t9    t10   1001 0011   -100   i1     i2    i3    i4   i1,i2   i1,i3   i1,i4   i3,i4   i2,i3   i2,i4
           original        VAE w/o { t4,t6}        IFM_DI w/o { t4,t6}              original     VAE w/o { t4,t6}      IFM_DI 5% w/o { t4,t6}



Figure 5: Details of the discrepancies on the number of duplicates (left) and the itemsets support (right)
between VAE and IFM-based techniques, without transactions 𝑡4 and 𝑡6 .


5. Conclusions
This paper has provided an overview about state-of-the-art approaches for synthetic transac-
tional data generation. A transaction has been modeled as a high-dimension sparse itemset,
that can be mapped into a binary vector, defined over the item space. The investigated algo-
rithms are (variants of the) Inverse Frequent Itemset Mining (IFM), and Probabilistic Generative
Models (PGMs). According to our analysis, the IFM approaches result to be extremely flexible
and understandable; they enable the control of the data generation procedure by the direct
identification of the discovery patterns to preserve. However, they proved to have extremely
onerous computational costs, making them not feasible in high-dimensional contexts. An
opposite conclusion has been obtained by analyzing PGMs: they are extremely fast and accurate,
but strongly lacking in control, flexibility and understandability. As future work, an interesting
research line is trying to investigate novel methodologies and techniques that are able to take
advantage of IFM and PGMs, by combining their strong points and mitigating their weakness.
Another promising research line is to apply the combination of the two approaches to NoSQL
applications by considering the extension of IFM that has been recently proposed in [12].


References
 [1] G. Manco, E. Ritacco, A. Rullo, D. Saccà, E. Serra, Machine learning methods for generating
     high dimensional discrete datasets, WIREs Data Mining Knowl. Discov. 12 (2022). URL:
     https://doi.org/10.1002/widm.1450. doi:10.1002/widm.1450.
 [2] C. P. Chen, C.-Y. Zhang, Data-intensive applications, challenges, techniques and tech-
     nologies: A survey on big data, Information Sciences 275 (2014) 314 – 347. doi:https:
     //doi.org/10.1016/j.ins.2014.01.015.
 [3] G. Weikum, Where’s the data in the big data wave?, 2013. ACM Sigmod Blog:
     http://wp.sigmod.org/?p=786.
 [4] R. Agrawal, T. Imieliński, A. Swami, Mining association rules between sets of items in
     large databases, in: Proceedings of the 1993 ACM SIGMOD International Conference on
     Management of Data, SIGMOD ’93, ACM, New York, NY, USA, 1993, pp. 207–216.
� [5] J. Han, H. Cheng, D. Xin, X. Yan, Frequent pattern mining: current status and fu-
     ture directions, Data Mining and Knowledge Discovery 15 (2007) 55–86. doi:10.1007/
     s10618-006-0059-1.
 [6] L. Cagliero, P. Garza, Itemset generalization with cardinality-based constraints, Informa-
     tion Sciences 244 (2013) 161 – 174. doi:https://doi.org/10.1016/j.ins.2013.05.
     008.
 [7] T. Mielikainen, On inverse frequent set mining, in: Proceedings of 2nd Workshop on
     Privacy Preserving Data Mining, PPDM ’03, IEEE Computer Society, Washington, DC,
     USA, 2003, pp. 18–23.
 [8] R. Agrawal, R. Srikant, Privacy-preserving data mining, in: Proceedings of the 2000 ACM
     SIGMOD International Conference on Management of Data, SIGMOD ’00, ACM, New
     York, NY, USA, 2000, pp. 439–450. doi:10.1145/342009.335438.
 [9] A. Guzzo, D. Saccà, E. Serra, An effective approach to inverse frequent set mining, in:
     Proceedings of the 2009 Ninth IEEE International Conference on Data Mining, ICDM ’09,
     IEEE Computer Society, Washington, DC, USA, 2009, pp. 806–811. doi:10.1109/ICDM.
     2009.123.
[10] A. Guzzo, L. Moccia, D. Saccà, E. Serra, Solving inverse frequent itemset mining with
     infrequency constraints via large-scale linear programs, ACM Transactions on Knowledge
     Discovery from Data 7 (2013) 18:1–18:39. doi:10.1145/2541268.2541271.
[11] D. Gunopulos, R. Khardon, H. Mannila, H. Toivonen, Data mining, hypergraph transversals,
     and machine learning, in: Proceedings of the 16-th ACM SIGMOD-SIGACT-SIGART
     Symposium on Principles of Database Systems, PODS ’97, 1997, pp. 209–216. doi:10.
     1145/263661.263684.
[12] D. Saccà, E. Serra, A. Rullo, Extending inverse frequent itemsets mining to generate realistic
     datasets: complexity, accuracy and emerging applications, Data Mining and Knowledge
     Discovery 33 (2019) 1736–1774. doi:10.1007/s10618-019-00643-1.
[13] P. Baldi, Autoencoders, unsupervised learning, and deep architectures, in: In Proceedings
     of the International Conference on Unsupervised and Transfer Learning workshop (UTLW),
     volume 27, PMLR, 2011, pp. 37–49.
[14] K. P. Murphy, Machine Learning: A Probabilistic Perspective, The MIT Press, 2012.
[15] D. M. Blei, A. Kucukelbir, J. D. McAuliffe, Variational inference: A review for statisti-
     cians, Journal of the American Statistical Association 112 (2017) 859–877. doi:10.1080/
     01621459.2017.1285773.
[16] D. Liang, R. G. Krishnan, M. Hoffman, T. Jebara, Variational autoencoders for collaborative
     filtering, in: Proceedings of the 2018 World Wide Web Conference, WWW ’18, 2018, pp.
     689–698.
[17] L. Theis, A. van den Oord, M. Bethge, A note on the evaluation of generative models,
     in: International Conference on Learning Representations, 2016. doi:10.48550/arXiv.
     1511.01844.
[18] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville,
     Y. Bengio, Generative adversarial nets, in: Advances in Neural Information Processing
     Systems, volume 27, 2014.
[19] M. Arjovsky, S. Chintala, L. Bottou, Wasserstein generative adversarial networks, in:
     Proceedings of the 34th International Conference on Machine Learning, 2017, pp. 214–223.
�