Vol-3184/paper3

From BITPlan ceur-ws Wiki
Revision as of 08:02, 6 March 2023 by Wf (talk | contribs)
Jump to navigation Jump to search

Paper

Paper
edit
description  
id  Vol-3184/paper3
wikidataid  →Q117040466
title  Enriching a Fashion Knowledge Graph from Product Textual Descriptions
pdfUrl  https://ceur-ws.org/Vol-3184/TEXT2KG_Paper_3.pdf
dblpUrl  
volume  →Vol-3184
session  →

Freitext

Enriching a Fashion Knowledge Graph from Product Textual Descriptions

load PDF

Enriching a Fashion Knowledge Graph from Product
Textual Descriptions
João Barroca1 , Abhishek Shivkumar1 , Beatriz Quintino Ferreira1 ,
Evgeny Sherkhonov1 and João Faria1
1
    Farfetch


                                         Abstract
                                         Knowledge Graphs offer a very useful and powerful structure for representing information, consequently,
                                         they have been adopted as the backbone for many applications in e-commerce scenarios. In this paper,
                                         we describe an application of existing techniques for enriching the large-scale Fashion Knowledge Graph
                                         (FKG) that we build at Farfetch. In particular, we apply techniques for named entity recognition (NER)
                                         and entity linking (EL) in order to extract and link rich metadata from product textual descriptions
                                         to entities in the FKG. Having a complete and enriched FKG as an e-commerce backbone can have a
                                         highly valuable impact on downstream applications such as search and recommendations. However,
                                         enriching a Knowledge Graph in the fashion domain has its own challenges. Data representation is
                                         different from a more generic KG, like Wikidata and Yago, as entities (e.g. product attributes) are too
                                         specific to the domain, and long textual descriptions are not readily available. Data itself is also scarce,
                                         as labelling datasets to train supervised models is a very laborious task. Even more, fashion products
                                         display a high variability and require an intricate ontology of attributes to link to. We use a transfer
                                         learning based approach to train an NER module on a small amount of manually labeled data, followed
                                         by an EL module that links the previously identified named entities to the appropriate entities within
                                         the FKG. Experiments using a pre-trained model show that it is possible to achieve 89.75% accuracy in
                                         NER even with a small manually labeled dataset. Moreover, the EL module, despite relying on simple
                                         rule-based or ML models (due to lack of training data), is able to link relevant attributes to products, thus
                                         automatically enriching the FKG.

                                         Keywords
                                         Fashion, Named Entity Recognition, Entity Linking, Knowledge Graph, E-commerce




1. Introduction
Knowledge Graphs (KGs) have proven to be essential in various domains including e-commerce
as they power many important business applications such as search and recommendation.
It is no exception for Farfetch1 , a global e-commerce platform for luxury fashion, where we
have recently embarked on the mission of building our own Fashion Knowledge Graph (FKG).


Text2KG 2022: International Workshop on Knowledge Graph Generation from Text, Co-located with the ESWC 2022,
May 05-30-2022, Crete, Hersonissos, Greece
$ joao.barroca@farfetch.com (J. Barroca); abhishek.shivkumar@farfetch.com (A. Shivkumar);
beatriz.quintino@farfetch.com (B. Q. Ferreira); evgeny.sherkhonov@farfetch.com (E. Sherkhonov);
joao.faria@farfetch.com (J. Faria)
� 0000-0002-6852-2716 (B. Q. Ferreira)
                                       © 2021 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
    CEUR
    Workshop
    Proceedings         CEUR Workshop Proceedings (CEUR-WS.org)
                  http://ceur-ws.org
                  ISSN 1613-0073




                  1
                      www.farfetch.com
�Figure 1: Example of a product that has no attributes specified but several attributes can be extracted
from the description such as material, fastening type, toe shape, and sole type.


Currently, the FKG contains entities like products, their attributes, such as color or material,
categories, product lines, brands, aesthetics, etc., that in total amount to a few billion triples.
   To keep the FKG up to date we have built scalable data pipelines that regularly update the
graph with thousands of products and metadata so it always contains fresh and relevant data.
The source data largely comes from structured data, e.g., internal data warehouses that are
populated from web forms for product information that are filled in by the back office or sellers.
As the latter process is manual and restricted to particular attribute types, it is often the case
that products miss a lot of important information which leads to an incomplete KG, and thus to
under-performing downstream applications. Therefore, as part of our KG construction pipeline,
we make use of Machine Learning models that extract product information from unstructured
data such as images and text. For instance, for the product depicted in Figure 1 the seller did
not specify any attribute information, but these can be extracted either from the image, such as
color information, or from the description, such as material, fastening type, toe type, and sole
type. To extract such information from product descriptions, as part of our FKG construction
pipeline we have built Named Entity Recognition (NER) and Entity Linking (EL) modules. The
NER module takes a product description as input, and tags sequences of strings, referred to as
mentions, with a possible attribute type. The EL module then links these tagged sequences to
the existing attribute values that are stored in the FKG. For example, for the product in Figure 1,
the NER module would tag the string "PVC" as material, then the EL module would link this
mention to :pvc, an entity in the FKG of type :Material.
   When building a KG for the e-commerce domain, we need to consider a few challenges.
First, there is the sheer complexity of the domain, with thousands to millions of products types,
organized into different taxonomy trees. In addition, there is a big variety across product types
regarding the attributes. Not only the attributes are different when considering different types of
products, but the attribute values vocabularies also vary across the different attributes. Together,
these challenges result in a lack of labeled data for the full taxonomy of products in our catalog.
   In this paper we demonstrate a simple approach for extracting mentions from product textual
descriptions, linking them to existing attribute values stored in the FKG, and consequently
�enriching the products with new attribute metadata. As main contributions of this work, we:

    • demonstrate how we leveraged transfer learning to train an NER model, based on BERT [1]
      Transformer model, for the fashion domain with very limited labeled data;
    • outline how simple rule-based models or traditional ML models using simple features can
      be used to build an EL system;
    • report some enrichment statistics after applying the NER and EL components, end to end,
      to the textual descriptions of a set of products in our catalogue.

   The rest of the paper is organized as follows. Section 2 gives an overview of some of the
relevant related works. In section 3, we describe the adopted methods, including a formalization
of the problem followed by a detailed explanation of each method. Section 4 lists all the
experiments done, including details of the small in-house labeled dataset used to train the
NER model. Lastly, section 5 reports the experimental results and analysis, followed by the
conclusions drawn in section 6.


2. Background and Related Work
As previously introduced, our approach can be interpreted as combining two main steps: first a
Named Entity Recognition (NER) task and, second, an Entity Linking task (EL). In the literature,
NER can be either seen as a component of EL to detect the mentions in the text, or can be also
handled independently from the EL system. At Farfetch, we had the manual resources available
to annotate a small NER dataset and so we decided to have the NER as a separate independent
component. In accordance, we will review the relevant state-of-the-art methods for NER and
EL separately.

2.1. Named Entity Recognition
The Named Entity Recognition (NER) task has been widely used in the information extraction
domain to identify entities, from unstructured text, and to classify them into pre-defined
categories (like names of persons, organizations, or attributes). [2] surveys long-established
NER approaches, prior to transformers.
   First approaches were commonly based on feature engineering and other supervised or
semi-supervised learning algorithms, including Support Vector Machines (SVMs), Conditional
Random Fields (CRFs) or Hidden Markov Models (HMMs) [2]. With the “deep learning” revolu-
tion, a leap in benchmark performance was achieved by both word-level and/or character-level
architectures based on applying a bidirectional-LSTM layer to word embeddings (namely
word2vec or glove) and then obtaining a final prediction from a CRF layer [3, 4]. LSTMs are
very suitable for capturing the semantics and context of the entities as they are able to handle
sparse context and have a sequential nature. On the other hand, LSTMs have been combined
with CRFs to enforce tagging consistency and to extract cohesive and meaningful entities classi-
fications. Attention mechanisms have also been applied (between the LSTM and CRF layers)
in order to increase NER model’s interpretability, i.e. to help explain the model’s decisions
by highlighting the importance of key concepts relative to their neighborhood context, as
�in [5]. Particularly, [5, 6] are framed in the e-commerce domain and also use product textual
information, such as title, descriptions, and bullet points. These works tackle the problem of
Attribute Value Extraction (getting all attribute values for a set of products) and model it as a
sequence tagging task, thus can be seen as a special case of NER. Marcelino et al., (2018) [7]
proposed a hierarchical deep learning natural language parser for fashion. Their approach not
only recognized fashion-domain entities but also exposed syntactic and morphologic insights.
   Transformers [8] have indelibly changed the NER landscape. Current state-of-the-art NER
methods are dominated by the Transformer architecture [8] which handles sequential input
data, like LSTMs, but solely relies on a self-attention mechanism. Such architecture brought
significant performance and efficiency improvements and Transformers blocks have been
extensively used for learning language representations applicable to many NLP downstream
tasks. Popular Transformer models pre-trained on very large corpora, that have been fine-tuned
to tackle the NER task, among others, include BERT [1], that introduces a “masked language
model” (MLM) pre-training objective, XLM-RoBERTa [9], which is also based on MLM but
outperforms multi-lingual BERT on NER benchmarks. Employing a pre-trained (self-supervised)
BERT-like model leverages the power of transfer learning in NLP, helping to solve the NER task
for specific domains where data is scarce, as we only need to fine-tune it on the specific domain
data.

2.2. Entity Linking
Entity Linking is the process of linking entity mentions appearing in unstructured text with
their corresponding entities in a knowledge base (KB). Although Entity Linking can be applied to
any kind of KB, in this review we will focus on methods that leverage the structured knowledge
that we can find in knowledge graphs.
   Apart from NER, an entity linking system has other three main components: candidate
generation, which is responsible for retrieving candidate entities for each mention from a KB;
candidate ranking/disambiguation, which is responsible for scoring and/or ranking the candidate
entities and choosing the one (or ones) to link; and, finally, unlinkable entities prediction (also
called NIL clustering) that deals with the mentions that were not linked to any entity.
   Before focusing on specific relevant methods found in the literature for each of these compo-
nents, and for the sake of completeness, we point the reader towards [10, 11, 12], where the
first two surveys comprehensively review EL methods based on Deep Learning, while the latter
presents an overview of recent advances in tasks for lifting Natural Language texts to KGs.
   Approaches to candidate entity generation have been primarily based on string matching
techniques, both for named dictionary based techniques and for surface form expansion [13]. In
such techniques, string matching is performed between the mention string (or the surface form
of the entity mention) and the name of the entity existing in the KB, usually using a named
dictionary. Moreover, some entity linking systems try to leverage the whole Web information
to identify candidate entities via Web search engines [13].
   The majority of proposed methods in the literature have either focused only on the candidate
ranking/disambiguation component or, more recently and with the advent of deep neural
networks, are end-to-end joint approaches. The latter approaches deal directly with an input text
and aim to extract all candidate mentions and link them to their corresponding entities in a KB.
�   Within disambiguation only approaches, Hoffart at al. [14] unifies prior approaches into
a comprehensive framework that combines three measures: popularity prior, similarity, and
coherence. In addition, it also introduces new measures for defining mention-entity similarity,
and a new algorithm for computing dense sub-graphs in a mention-entity graph, which produces
high quality mention-entity mappings. However, the main contribution of [14] is one of the best
known benchmarks in the entity disambiguation task: the AIDA CoNLL-YAGO - a new dataset
based on CoNLL 2003, in which they manually annotated all proper nouns with corresponding
entities in YAGO2.
   Recently, transformers have been widespreadly used in disambiguation tasks, leading to per-
formance leaps on multiple benchmarks. Specifically, Yamada et al. [15] introduce a confidence-
order model that achieves top results in several benchmarks, by making use of the recent
pre-trained BERT model. In this work, the masked language model task is tackled as a masked
entity prediction task. Considering the context provided by KGs, Mulang et al. [16] demon-
strate that pre-trained transformer-based models, although powerful, are limited to capturing
context available purely on the texts concerning the original training corpus. They observe
that adding an extra, task-specific, KG context improves the performance of Candidate Entity
Disambiguation methods, leading to a new best performance for AIDA-CoNLL dataset.
   As for end-to-end approaches, relevant NN-based end-to-end linking models include: Ment-
norm [17], Kolitsas et al. [18], NCEL [19], and Martins et al. [20]. Ment-norm [17] achieved
the best end-to-end results followed by Kolitsas et al. [18]. Kolitsas et al. [18] use a bi-LSTM
based model for mention detection and computes the similarity between the entity mention
embedding and a set of predefined entity candidates. The authors demonstrate that engineered
features are almost unnecessary when using end-to-end approaches. Their model reaches
SOTA results on the AIDA/CoNLL dataset (comparing with the other end-to-end methods)
and, when combined with Stanford NER, it generalizes well to other datasets with different
characteristics. Ment-norm [17] is an end-to-end system for NEL that considers relations
as latent. Representation learning was used to learn relation embeddings, eliminating the
need for extensive feature engineering. Martins et al. (2019) [20] also explore joint learning
of Named Entity Recognition (NER) and EL showing that the two tasks benefit from joint
training. More recently, Cao et al. [21] propose GENRE, a novel paradigm to address entity
retrieval: it generates entity names autoregressively (leveraging pre-trained autoregressive
language model BART and tackling EL as sequence-to-sequence problem). m-GENRE [22]
extends GENRE [21] to multi-language. In [23], Cao et al., make an improvement over the
previous study (GENRE[21]) that boosts performance as it parallelizes autoregressive linking
across all potential mentions and relies on a shallow and efficient decoder. Moreover, they add
a new discriminative component that optimizes generators ranking. Finally, Ravi et al. [24]
propose CHOLAN, a modular approach to target end-to-end entity linking. CHOLAN consists
of a pipeline of two transformer-based models integrated sequentially to accomplish the EL
task. The first transformer model performs Named Entity Recognition in a given text and the
second performs Named Entity Disambiguation (using the Wikipedia pages as entities external
context).
   Comparing disambiguation-only with end-to-end methods, taking into account benchmark
results reported in the previously mentioned works, we observe that disambiguation-only
methods usually outperform end-to-end methods in the entity disambiguation task. However,
�the recent GENRE [21], has competitive results for entity disambiguation and achieved SOTA
in the entity linking task. Nevertheless, end-to-end models are usually based on deep networks
which need a huge amount of training data to generate robust models that perform well. In
fact, [18] has shown that the accuracy of end-to-end models drops dramatically when trained on
or applied to small datasets. This becomes a serious limitation especially when we’re performing
EL in specific (small) domains, as is the case of in work.


3. Method
Recently at Farfetch we embarked on the journey of building our own Fashion Knowledge Graph
(FKG). The purpose of the FKG is to model and structure our knowledge about fashion and
unify information about the main entities such as products, brands, aesthetics, etc. Such unified
information can be effectively used in various applications such as search and recommendations.
   The FKG is modelled as a set of RDF2 triples, i.e., tuples of the form (𝑠, 𝑝, 𝑜), where 𝑠 and
𝑝, called the subject and the property, are an IRI, and 𝑜, called the object, is an IRI, blank node,
or a literal. In the FKG, entities can be instances of different classes, such as :Product or
:Brand, and be connected via properties such as :hasBrand and :hasAesthetic. Since in
this paper we are interested in enriching the FKG with the product attributes extracted from the
textual descriptions, we therefore enrich only a single property :hasAttribute that connects
products to its attributes in the FKG. Rest of the data for FKG, on the other hand, comes from
other internal or external data sources and enrichers.
   In what follows, we outline the methods involved in extracting the attribute mentions from
the textual product descriptions and linking them to the attribute in the FKG. Attribute mentions
are the sequence of strings extracted from the textual product descriptions that describe a
possible attribute value. First, we define and formalize our task. Next, we detail our method for
detecting attribute mentions in unstructured text. Finally, we explain how these mentions can
be linked to the attribute values presented in the FKG.

3.1. Task Definition
Given a textual product description, our objective is to extract relevant product attributes. The
relevant attributes are the attributes presented in our Fashion Knowledge Graph.
   We formulate this task as a two step process. The first step is a Named Entity Recognition
(NER) component, which detects attribute mentions in the unstructured text. In addition, NER
also detects the type of attribute associated with each mention. Formally, consider a product 𝑃
and its textual description 𝑇 𝑃 = (𝑡𝑃1 , 𝑡𝑃2 , ..., 𝑡𝑃𝑛 ) where 𝑡𝑃𝑖 refers to the 𝑖𝑡ℎ token of the product
description. The named entity recognition task requires to detect a set 𝑀 𝑃 = {𝑚𝑃1 , 𝑚𝑃2 , ..., 𝑚𝑃𝑘 },
𝑘 ≤ 𝑛, such that every 𝑚𝑃𝑗 is a pair of mention and mention type. In the following we often
use the notions of mentions and mention pairs interchangeably. For the example in Figure 1,
the NER detects the following attribute mention and mention type pairs: ("PVC", material),
("lace-up front fastening", closure type), ("round toe", toe shape), and ("ridged
rubber sole", sole type).

    2
        https://www.w3.org/TR/PR-rdf-syntax/
�Figure 2: A sample product description showing the annotated entities.


   The second step is an Entity Linking (EL) component, in which the detected attribute mention
pairs are linked to a set of entities from the FKG. Formally, consider the previous detected
attribute mentions 𝑀 𝑃 . In addition, we are given a set of attribute entity and type pairs
𝐸 = {𝑒1 , 𝑒2 , ..., 𝑒𝑙 } extracted from the FKG. The entity linking task is the task of linking each
mention 𝑚𝑃𝑗 ∈ 𝑀 𝑃 to a unique entity 𝑒𝑃𝑖 ∈ 𝐸 ∪ {NIL}, where NIL is a special null symbol
to denote that no entity is linked. As an example, consider the following set of entity and
entity type pairs, corresponding to specific product attributes, retrieved from the FKG: {(:pvc,
:Material), (:cotton, :Material), (:round, :Toe Shape)}. The detected mention pairs
from the previous example would be linked with to following entities:
    • ("PVC", material) → (:pvc, :Material)
    • ("lace-up front fastening", closure type) → NIL
    • ("round toe", toe shape) → (:round, :Toe Shape)
    • ("ridged rubber sole", sole type) → NIL

3.2. Named Entity Recognition
Due to the lack of labeled NER data within the Fashion domain, we make use of a pre-trained
Transformer language model that has already been trained on a large amount of unstructured
text. We take the weights from the pre-trained BERT-base-cased model [1] and fine-tune it on
top of a small set of manually labeled NER dataset. We used an in-house built user-interface tool
to manually label 900 descriptions of products from dresses category across 11 classes. We used
the standard BIO 3 (short for beginning, inside, outside) notation for labelling the named entities,
a common tagging format for tagging tokens in a chunking task in computational linguistics.
The B-tag is used to indicate that the token is the beginning of a named entity phrase, while the
I-tag is used for tokens inside the phrase, and the O-tag is used for tokens that do not belong to
any named entity chunk. A sample product description with labeled named entities is shown in
Figure 2.

3.3. Entity Linking
As explained in Section 2, traditional entity linking (EL) systems are composed of three main
components: candidate entity generation, candidate entity disambiguation and NIL clustering. In
our solution, we only focus on the first 2 components. Figure 3 describes the full end-to-end
system.
   3
       https://en.wikipedia.org/wiki/Inside-outside-beginning_(tagging)
�Figure 3: Full end to end system. Given a product textual description, the NER component detects
attribute mentions and their corresponding types in the text. Next, for each attribute mention, the
candidate entity generator retrieves a set of candidate attribute entities. Lastly, the candidate entity
disambiguation predicts a score for each candidate, and decides which candidates should be linked or not.
In the example described in the diagram, one of the attribute mentions detected is "white", which is
tagged as color. Given this mention, the following candidate entities are retrieved: (:white, :Color),
(:off-white, :Color), and (:white gold, :Material). Since the entity (:white, :Color) is the
one with higher score, and because the score is larger than a defined threshold, the entity is linked.


3.3.1. Candidate Entity Generation
Candidate entity generation can be reformulated as an information retrieval (IR) problem: given
a query, i.e. an attribute mention, we want to retrieve a set of results relevant to that query, i.e.
a set of candidate entities. Of the variety of IR techniques, the most widely used are based on
the BM25 [25] algorithm. BM25, where BM is an abbreviation of best matching, is a non-linear
retrieval function that combines three key document attributes: term frequency, document
frequency, and document length. There are a lot of available search engines that allow us to
search for content based on inverted indexes, using the BM25 algorithm as a ranking function.
   The attribute values in the FKG are represented as entities with labels as string literals. In
order to be able to search over the attribute values in the FKG, we first create an inverted index
for all their labels, and then use that index to perform full-text search using the BM25 function
to rank the results.

3.3.2. Candidate Entity Disambiguation
Candidate entity disambiguation is formulated as a binary classification problem. Given a pair
of an attribute mention and a candidate attribute entity, we predict if the candidate should
be linked or not. We highlight, at this point, that we lack the training data needed to train
�Table 1
exact-match and sub-match examples.

                            mention             entity   exact-match   sub-match

                          bright pink     bright pink       link         link
                          bright pink         pink        no_link        link
                              pink        bright pink     no_link      no_link
                          bright pink      light pink     no_link      no_link



more complex solutions based on Neural Networks. Therefore, we test two types of models:
rule-based models and traditional ML models.
   The rule-based models are based on domain-specific heuristics. First, we build a consistent
text-preprocessing pipeline that enforces lower-casing, text normalization (including accents
normalization and special characters removal), and stemming. Next, we create a type mapper
which only allows certain types of entities to be linked with certain types of mentions. Finally,
we build two different rule-based models. One of them performs an exact-match between
the mention and the candidate entity tokens. The other verifies if the candidate entity tokens
are contained within the mention tokens, a matching that we name as sub-match. We do this
because we want to link to a broader entity, but never to a narrower entity in order to guarantee
the validity of what is linked. Table 1 exemplifies some matching cases.
   The traditional ML models we explore are binary classifiers that predict a binary target for
each mention-candidate pair. Since these models need numeric features as input, we build a set
of features for each mention-candidate pair that are used by the binary classifier to make its
prediction. These features can be split into two main types:

    • string similarity features: as the name suggests, these features use string similarity mea-
      sures. In particular, we use jaro-winkler4 to generate the similarity scores between the
      mention and the candidate entity tokens.
    • semantic similarity features: these features use vector representations (embeddings) for
      the mention and candidate entity, and then compute the cosine similarity between these
      representations. In order to generate specific embeddings for the fashion domain, we
      used a pre-trained BERT-based language model, fine-tuned on 236K textual descriptions
      from our product catalog, using Masked Language Modeling as in [1].

    In order to keep the EL independent of the attribute types, not only do we generate the
previous features using the mention and entity names, but we also use their types. This is in
contrast with the rule-based models, in which we define a type mapper heuristic based on our
domain expertise.
    During inference, instead of predicting the binary label directly, we compute a linking score
(i.e. the score associated with the positive target) and then rank all the candidate entities, for
the same mention, by the linking score. Then, if the score of the top-rank candidate is larger
than a specific threshold (that can be tuned during training), the attribute entity is linked.

   4
       https://pypi.org/project/jaro-winkler/
�Table 2
Number of instances of named entities across the different entity types for both train and validation sets
              Detail   Color   Shape   Neckline   Pattern   Sleeve Type   Hem   Material   Fastening   Collar   Pockets
 Train         896     797      759      571       403         699        648     491        352        74        12
 Validation    251     203      176      138       74          169        167     116         86        25        8



4. Experiments
In this section, we outline the experiments conducted for the different components. In the first
experiment, we fine-tune a BERT model in an NER task. In the second experiment, we train
a few binary classifier models and compare them to our rule-based models, for the candidate
entity disambiguation task. Both experiments use a small in-house dataset composed of textual
product descriptions from the fashion domain.

4.1. Named Entity Recognition
As the world of fashion is a very complex domain with lots of categories and sub-categories
arranged in a hierarchical manner, we focused only on Dresses category for our experiments. We
fine-tuned a pre-trained Transformer BERT-base-cased model [1] for NER using 900 manually
labeled product descriptions across 11 different tags, viz. shape, hem, color, collar, material,
fastening, pockets, neckline, sleeve_type, pattern, and detail. The pre-trained model was trained
on Wikipedia5 and the BookCorpus6 datasets using Masked Language Modeling and Next
Sentence Prediction tasks [1]. There were 5702 and 1413 instances of named entities in the
training and validation set, respectively, across different tags as shown in Table 2. Here, we
used the Hugging Face library7 to fine-tune the pre-trained BERT model with the last linear
layer on top configured to classify the named entities within our dataset. The model consisted
of 12 self-attention heads and 12 hidden layers. Each hidden layer had 768 units and there
were 512 maximum positional embeddings. The hidden layer units used gelu activation units
with a dropout probability of 0.1. We also used the BERT pre-trained tokenizer consisting
of a vocabulary size of 28996 tokens. Hugging Face enabled us to download the model with
pre-trained weights.
   We fine-tuned the above model on 900 labeled samples, split into 720 for training and 180 for
validation. Even though the dataset is small in size, we saw that having access to a pre-trained
BERT-base-cased model provides us with the high accuracy necessary for the entity linking
step.

4.2. Entity Linking
The first component to experiment with was the candidate entity generation. As explained
in Section 3.3, we use full-text search over an inverted index of attribute values. In order to
increase the recall set of the candidate attribute entities retrieved, we use fuzzy matching. We
    5
      https://huggingface.co/datasets/wikipedia
    6
      https://huggingface.co/datasets/bookcorpus
    7
      www.huggingface.com
�Table 3
Number of instances of named entities across the different entity types for both train and validation sets
 Material      Pattern   Neckline   Sleeve Type   Detail   Closure Type   Length   Color   Sleeve Length   Shape
    444          397       280         248         179         174         137     134          91          70



experimented with fuzziness scores in the range of [0.5, 1.0], using increments of 0.1, and
manually tested a small list of attributes that should be retrieved for specific queries. We ended
up choosing a fuzziness score of 0.7, which means that we match all terms having 70% matching
characters with the query.
  Next, we trained and compared different models for the candidate entity disambiguation
component. Since we needed labeled data to train (and test) the models, we manually annotated
32 instances, randomly sampled from the dataset used to train the NER model. Each instance
contains a product textual description and the corresponding attribute mentions that should
be recognized from the text. For each mention, we generate the candidate entities using the
candidate entity generation module. We ended up with 2528 samples, where each sample is
composed of a product textual description, an attribute mention, and an attribute candidate
entity. Lastly, we manually labeled each sample using binary labels, with the positive samples
representing the candidate attribute entities that should be linked, yielding 181 (7.2%) positive
samples and 2347 negative samples. The final dataset contains 132 unique attribute mentions,
covering the 11 unique types described in Table 2, and 522 unique attribute entities covering
33 unique types of attributes (the top 10 attribute types with more samples are represented in
Table 3). We used a train/test split ratio of 9:1.
  Regarding the candidate entity disambiguation models, we used the two rule-based models
described in Section 3.3 as baselines, and trained 3 different binary classifiers: a logistic
regression classifier, a linear support vector classifier (SVC), and a random
forest classifier. As metrics, we used the standard metrics for binary classification:
precision, recall and f1-score. We do not report the accuracy score due to the high unbalanced
nature of the dataset.


5. Results
In this section, we report the accuracy numbers for the NER model and the results from the
classification model for the candidate entity disambiguation model. In addition, we also report a
few enrichment statistics using the full system end to end.

5.1. Named Entity Recognition
To evaluate the performance of the fine-tuned BERT-base-cased model, we used the seqeval8
framework in Python. The fine-tuned BERT model achieved an accuracy of 89.75% on the
validation set. Due to the small size of the dataset, we had to stop the training of the model
early enough to prevent overfitting.
    8
        https://github.com/chakki-works/seqeval
�Figure 4: Visualization of the self-attention map for a sample product description.


Table 4
Binary classification results on the test set, for candidate entity disambiguation. The best results are
highlighted in bold, with the second best in underline.
                      Model                        Precision   Recall   F1-Score
                      exact-match                    1.00       0.56      0.72
                      sub-match                      0.78       0.72      0.75
                      logistic-regression            0.94       0.68      0.79
                      svc                            0.89       0.68      0.77
                      random-forest                  0.91       0.80      0.85


  We also visualized the weights of the self-attention layer showing the importance between the
tokens in the input text. For example, as shown in Figure 4, the token red has a high correlation
with tokens including fashionable and color. This way, we believe that the model learned an
inner representation of the fashion domain language that can then be used to extract features
useful for downstream tasks.

5.2. Entity Linking
We report the binary classification results on the test set, for the candidate entity disambiguation
task, in Table 4. As expected, the exact-match rule-based model has a perfect precision.
Although there may be a few ambiguous cases in which there is an exact match with multiple
attributes, we are constraining the linking based on the mention and entity types using a
type mapper heuristic, and therefore disambiguating such cases. For example, consider the
attribute mention "long", labeled with the type sleeve_type by NER (recall Table 2). Two
exact-matched candidate entities are (:long, :Length) and (:long, :Sleeve Length). Due
to the type mapper heuristic, only the (:long, :Sleeve Length) attribute entity is linked.
The high precision of the exact-match rule-based model comes with the trade-off of a low
recall (only 56% of the entities are linked).
   In contrast, due to its less stringent matching mechanism, the sub-match rule-based
�Table 5
Top 2 mentions without linked entities
                   Mention                "fitted waist"       "straight hem"
                   Mention Type                shape                 hem
                   Number of Products             63                   31


model has a higher recall (0.72), while maintaining a decent precision (0.783), and thereby
achieving an f1-score of 0.75.
  Regarding the traditional ML binary classifiers, during training, we estimated the score
thresholds for each model in order to keep a minimum precision of 0.9, using cross-validation
experiments. The results reported in Table 4 correspond to the results on the test set, using
the score threshold estimated during training. Although the logistic regression has the
highest precision (0.94), both the support vector classifier and the random forest
models keep the minimum precision requirement of 0.9. The random forest classifier is the
model that yields the best overall results, with the highest f1-score (0.85), and therefore it is the
model implemented in the candidate entity disambiguation component.

5.3. End to End System
The previous experiments and results describe the performance of both the NER and EL compo-
nents individually. In order to estimate how the end to end system is performing, we applied
the full system to a set of 650 product textual descriptions, extracted from the dresses category.
On average, for each product, we detect 7.79 attribute mentions, we generate 12.8 candidate
entities for each mention, and end-up linking 5.68 attribute entities. From all the entities linked,
52% are exact matches.
   In addition, there are an average of 2.63 attribute mentions that are not linked per product.
Table 5 reports the top 2 mentions without any links, that are detected in a large number of
products. The "straight hem" mention represents a limitation of EL, which could easily be
solved by improving the text-preprocessing, since we can find the attribute entity (:straight
hemline, :Hemline Style) among the candidates. However, regarding the "fitted waist"
mention, there are no relevant candidates. Since this mention occurs with high frequency in the
products textual descriptions, it is a good indicator that we may need to create a corresponding
new attribute entity in the FKG.
   Although the previous statistics demonstrate that we are enriching the products with an
average of 5.68 attributes, and the EL experimental results report a linking precision of 0.9,
we are still not completely certain about the correctness of the full enrichment, due to error
propagation from the NER component to the downstream modules. Consider the following
product textual description excerpt: "a bright blue dress, featuring a". A possible error would be
for the NER to detect ("blue", color), instead of ("bright blue", color). This error would
lead to the linking of the attribute (:Blue, :Color), instead of (:Bright Blue, :Color), in
the EL component. Although not critical, such error propagation raises the need for improving
the robustness of EL.
�6. Conclusion
In this paper, we have described how we applied existing methods to tackle the automatic
enrichment of our KG within the fashion domain. We tackled the NER and the EL tasks and,
despite the challenges brought by the lack of data and the specificity of the domain, we were
able to augment our FKG, which is pivotal for the success of downstream tasks. In particular,
we have shown how we leveraged transfer learning approaches to fine-tune an NER model to
the fashion domain using a small dataset. In addition, by using traditional ML models with
simple features, we also built an EL component using a few labeled samples.
   We are also confident that the embeddings obtained by fine-tuning the BERT-based model
on the fashion domain capture the context and semantics specific to the domain. Hence, these
domain-specific embeddings, which are a by-product of our enriching method, may enable
several other downstream tasks.
   However, our solution still presents several limitations. One of them is that the errors from
the NER model may propagate into the EL step as a result of having a 2-step pipeline. This could
be tackled by introducing further robustness matching steps. Having a jointly optimized process
could also present advantages worth exploring. Nevertheless, such an approach requires the
construction of a new dedicated dataset. So far, because of the efforts required for annotating
data, our focus has been only on a subset of attributes for the Dresses category, and we must scale
this process to other attribute types and product categories in order to cover a larger set of our
catalog. Therefore, we are focusing on defining the high-impact attributes for the business, and
developing specific individual models targeting them, since it’s crucial to guarantee high-quality
results. However, for the remaining attributes, we may experiment with recent attribute value
extraction methods, as in [6], that incorporate the attribute types directly as features and thereby
are able to scale to thousands of attribute types, while requiring few annotated data points for
each type. Furthermore, we feel there is still scope for improvement on entity disambiguation,
across multiple levels, namely on text pre-processing, better feature engineering by generating
richer features, and increasing model complexity.
   Besides the above limitations, and as an additional direction for future work, we feel we can
take advantage of the NIL clustering to detect new attributes in order to automatically update
the ever-growing and ever-changing taxonomy of FKG.


Acknowledgments
The authors would like to thank all team members of the FKS team at Farfetch. Their support
was crucial for the conducted research.


References
 [1] J. Devlin, M. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional trans-
     formers for language understanding, arXiv preprint arXiv:1810.04805 (2018).
 [2] V. Yadav, S. Bethard, A survey on recent advances in named entity recognition from
�     deep learning models, CoRR abs/1910.11470 (2019). URL: http://arxiv.org/abs/1910.11470.
     arXiv:1910.11470.
 [3] Z. Huang, W. Xu, K. Yu, Bidirectional lstm-crf models for sequence tagging, CoRR
     abs/1508.01991 (2015).
 [4] G. Lample, M. Ballesteros, S. Subramanian, K. Kawakami, C. Dyer, Neural architectures for
     named entity recognition, in: Proceedings of the 2016 Conference of the North American
     Chapter of the Association for Computational Linguistics: Human Language Technologies,
     2016, pp. 260–270.
 [5] G. Zheng, S. Mukherjee, X. L. Dong, F. Li, Opentag: Open attribute value extraction from
     product profiles, in: Proceedings of the 24th ACM SIGKDD International Conference
     on Knowledge Discovery & Data Mining, KDD ’18, 2018, p. 1049–1058. doi:10.1145/
     3219819.3219839.
 [6] H. Xu, W. Wang, X. Mao, X. Jiang, M. Lan, Scaling up open tagging from tens to thousands:
     Comprehension empowered attribute value extraction from product title, in: Proceedings
     of the 57th Conference of the Association for Computational Linguistics, ACL 2019, 2019,
     pp. 5214–5223.
 [7] J. Marcelino, J. Faria, L. Baía, R. G. Sousa, A hierarchical deep learning natural language
     parser for fashion, arXiv preprint arXiv:1806.09511 (2018).
 [8] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. Gomez, L. Kaiser, I. Polo-
     sukhin, Attention is all you need, in: Advances in Neural Information Process-
     ing Systems, volume 30, 2017. URL: https://proceedings.neurips.cc/paper/2017/file/
     3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf.
 [9] A. Conneau, K. Khandelwal, N. Goyal, V. Chaudhary, G. Wenzek, F. Guzmán, E. Grave,
     M. Ott, L. Zettlemoyer, V. Stoyanov, Unsupervised cross-lingual representation learning at
     scale, in: Proceedings of the 58th Annual Meeting of the Association for Computational
     Linguistics, 2020, pp. 8440–8451.
[10] Ö. Sevgili, A. Shelmanov, M. Y. Arkhipov, A. Panchenko, C. Biemann, Neural entity
     linking: A survey of models based on deep learning, CoRR abs/2006.00575 (2020). URL:
     https://arxiv.org/abs/2006.00575. arXiv:2006.00575.
[11] W. Shen, Y. Li, Y. Liu, J. Han, J. Wang, X. Yuan, Entity linking meets deep learning:
     Techniques and solutions, CoRR abs/2109.12520 (2021). URL: https://arxiv.org/abs/2109.
     12520. arXiv:2109.12520.
[12] T. Al-Moslmi, M. O. Gallofre, A. L. Opdahl, C. Veres, Named entity extraction for knowledge
     graphs: A literature overview, IEEE Access 8 (2020) 32862–32881. doi:10.1109/ACCESS.
     2020.2973928.
[13] W. Shen, J. Wang, J. Han, Entity linking with a knowledge base: Issues, techniques, and
     solutions, IEEE Transactions on Knowledge and Data Engineering 27 (2015) 443–460.
     doi:10.1109/TKDE.2014.2327028.
[14] J. Hoffart, M. A. Yosef, I. Bordino, H. Fürstenau, M. Pinkal, M. Spaniol, B. Taneva, S. Thater,
     G. Weikum, Robust disambiguation of named entities in text, in: Proceedings of the
     2011 Conference on Empirical Methods in Natural Language Processing, Association
     for Computational Linguistics, Edinburgh, Scotland, UK., 2011, pp. 782–792. URL: https:
     //aclanthology.org/D11-1072.
[15] I. Yamada, H. Shindo, Pre-training of deep contextualized embeddings of words and entities
�     for named entity disambiguation, CoRR abs/1909.00426 (2019). URL: http://arxiv.org/abs/
     1909.00426. arXiv:1909.00426.
[16] I. O. Mulang, K. Singh, C. Prabhu, A. Nadgeri, J. Hoffart, J. Lehmann, Evaluating the impact
     of knowledge graph context on entity disambiguation models, CoRR abs/2008.05190 (2020).
     URL: https://arxiv.org/abs/2008.05190. arXiv:2008.05190.
[17] P. Le, I. Titov, Improving entity linking by modeling latent relations between mentions,
     in: Proc. 56th Annu. Meeting Assoc. Comput. Linguistics, 2018, p. 1595–1604.
[18] N. Kolitsas, O. E. Ganea, T. Hofmann, End-to-end neural entity linking, in: Proc. 22nd
     Conf. Comput. Natural Lang. Learn., 2018, p. 519–529.
[19] Y. Cao, L. Hou, J. Li, Z. Liu, Neural collective entity linking, in: Proc. 27th Int. Conf.
     Comput. Linguistics, 2018, p. 675–686.
[20] P. H. Martins, Z. Marinho, A. F. T. Martins, Joint learning of named entity recognition
     and entity linking, in: Proc. 57th Annu. Meeting Assoc. Comput. Linguistics, Student Res.
     Workshop, 2019, p. 190–196.
[21] N. D. Cao, G. Izacard, S. Riedel, F. Petroni, Autoregressive entity retrieval, CoRR
     abs/2010.00904 (2020). URL: https://arxiv.org/abs/2010.00904. arXiv:2010.00904.
[22] N. D. Cao, L. Wu, K. Popat, M. Artetxe, N. Goyal, M. Plekhanov, L. Zettlemoyer, N. Cancedda,
     S. Riedel, F. Petroni, Multilingual autoregressive entity linking, CoRR abs/2103.12528
     (2021). URL: https://arxiv.org/abs/2103.12528. arXiv:2103.12528.
[23] N. D. Cao, W. Aziz, I. Titov, Highly parallel autoregressive entity linking with discrim-
     inative correction, CoRR abs/2109.03792 (2021). URL: https://arxiv.org/abs/2109.03792.
     arXiv:2109.03792.
[24] M. Ravi, K. Singh, I. O. Mulang, S. Shekarpour, J. Hoffart, J. Lehmann, CHOLAN: A modular
     approach for neural entity linking on wikipedia and wikidata, CoRR abs/2101.09969 (2021).
     URL: https://arxiv.org/abs/2101.09969. arXiv:2101.09969.
[25] S. E. Robertson, S. Walker, Some simple effective approximations to the 2-poisson model
     for probabilistic weighted retrieval, in: Proceedings of the 17th Annual International
     ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’94,
     Springer-Verlag, Berlin, Heidelberg, 1994, p. 232–241.