<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>http://ceur-ws.bitplan.com/index.php?action=history&amp;feed=atom&amp;title=Vol-3194%2Fpaper16</id>
	<title>Vol-3194/paper16 - Revision history</title>
	<link rel="self" type="application/atom+xml" href="http://ceur-ws.bitplan.com/index.php?action=history&amp;feed=atom&amp;title=Vol-3194%2Fpaper16"/>
	<link rel="alternate" type="text/html" href="http://ceur-ws.bitplan.com/index.php?title=Vol-3194/paper16&amp;action=history"/>
	<updated>2026-04-04T21:07:59Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.35.5</generator>
	<entry>
		<id>http://ceur-ws.bitplan.com/index.php?title=Vol-3194/paper16&amp;diff=1593&amp;oldid=prev</id>
		<title>Wf: modified through wikirestore by wf</title>
		<link rel="alternate" type="text/html" href="http://ceur-ws.bitplan.com/index.php?title=Vol-3194/paper16&amp;diff=1593&amp;oldid=prev"/>
		<updated>2023-03-30T16:51:22Z</updated>

		<summary type="html">&lt;p&gt;modified through wikirestore by wf&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;=Paper=&lt;br /&gt;
{{Paper&lt;br /&gt;
|id=Vol-3194/paper16&lt;br /&gt;
|storemode=property&lt;br /&gt;
|title= Exploiting Curated Databases to Train Relation Extraction Models for Gene-Disease Associations&lt;br /&gt;
|pdfUrl=https://ceur-ws.org/Vol-3194/paper16.pdf&lt;br /&gt;
|volume=Vol-3194&lt;br /&gt;
|authors=Stefano Marchesin,Gianmaria Silvello&lt;br /&gt;
|dblpUrl=https://dblp.org/rec/conf/sebd/0001S22&lt;br /&gt;
}}&lt;br /&gt;
== Exploiting Curated Databases to Train Relation Extraction Models for Gene-Disease Associations==&lt;br /&gt;
&amp;lt;pdf width=&amp;quot;1500px&amp;quot;&amp;gt;https://ceur-ws.org/Vol-3194/paper16.pdf&amp;lt;/pdf&amp;gt;&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
Exploiting Curated Databases to Train Relation&lt;br /&gt;
Extraction Models for Gene-Disease Associations*&lt;br /&gt;
(Discussion Paper)&lt;br /&gt;
&lt;br /&gt;
Stefano Marchesina , Gianmaria Silvelloa&lt;br /&gt;
a&lt;br /&gt;
    Dipartimento di Ingegneria dell’Informazione, Università degli Studi di Padova, Via Gradenigo 6/b, 35131, Padova, Italy&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
                                         Abstract&lt;br /&gt;
                                         Databases are pivotal to advancing biomedical science. Nevertheless, most of them are populated and&lt;br /&gt;
                                         updated by human experts with a great deal of effort. Biomedical Relation Extraction (BioRE) aims to&lt;br /&gt;
                                         shift these expensive and time-consuming processes to machines. Among its different applications, the&lt;br /&gt;
                                         discovery of Gene-Disease Associations (GDAs) is one of the most pressing challenges. Despite this,&lt;br /&gt;
                                         few resources have been devoted to training – and evaluating – models for GDA extraction. Besides,&lt;br /&gt;
                                         such resources are limited in size, preventing models from scaling effectively to large amounts of data.&lt;br /&gt;
                                         To overcome this limitation, we have exploited the DisGeNET database to build a large-scale, semi-&lt;br /&gt;
                                         automatically annotated dataset for GDA extraction: TBGA. TBGA is generated from more than 700K&lt;br /&gt;
                                         publications and consists of over 200K instances and 100K gene-disease pairs. We have evaluated state-&lt;br /&gt;
                                         of-the-art models for GDA extraction on TBGA, showing that it is a challenging dataset for the task. The&lt;br /&gt;
                                         dataset and models are publicly available to foster the development of state-of-the-art BioRE models for&lt;br /&gt;
                                         GDA extraction.&lt;br /&gt;
&lt;br /&gt;
                                         Keywords&lt;br /&gt;
                                         Weak Supervision, Relation Extraction, Gene-Disease Association&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
1. Introduction&lt;br /&gt;
Curated databases, such as UniProt [2], DrugBank [3], or CTD [4], are pivotal to the develop-&lt;br /&gt;
ment of biomedical science. Such databases are usually populated and updated with a great deal&lt;br /&gt;
of effort by human experts [5], thus slowing down the biological knowledge discovery process.&lt;br /&gt;
To overcome this limitation, the Biomedical Information Extraction (BioIE) field aims to shift&lt;br /&gt;
population and curation processes to machines by developing effective computational tools that&lt;br /&gt;
automatically extract meaningful facts from the vast unstructured scientific literature [6, 7, 8].&lt;br /&gt;
Once extracted, machine-readable facts can be fed to downstream tasks to ease biological knowl-&lt;br /&gt;
edge discovery. Among the various tasks, the discovery of Gene-Disease Associations (GDAs)&lt;br /&gt;
is one of the most pressing challenges to advance precision medicine and drug discovery [9],&lt;br /&gt;
as it helps to understand the genetic causes of diseases [10]. Thus, the automatic extraction&lt;br /&gt;
&lt;br /&gt;
                 * The full paper has been originally published in BMC Bioinformatics [1]&lt;br /&gt;
SEBD 2022: The 30th Italian Symposium on Advanced Database Systems (SEBD 2022), June 19–22, 2022, Pisa, Italy&lt;br /&gt;
Envelope-Open stefano.marchesin@unipd.it (S. Marchesin); gianmaria.silvello@unipd.it (G. Silvello)&lt;br /&gt;
Orcid 0000-0003-0362-5893 (S. Marchesin); 0000-0003-4970-4554 (G. Silvello)&lt;br /&gt;
                                       © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).&lt;br /&gt;
    CEUR&lt;br /&gt;
    Workshop&lt;br /&gt;
    Proceedings&lt;br /&gt;
                  http://ceur-ws.org&lt;br /&gt;
                  ISSN 1613-0073&lt;br /&gt;
                                       CEUR Workshop Proceedings (CEUR-WS.org)&lt;br /&gt;
�and curation of GDAs is key to advance precision medicine research and provide knowledge to&lt;br /&gt;
assist disease diagnostics, drug discovery, and therapeutic decision-making.&lt;br /&gt;
   Most datasets used to train and evaluate Relation Extraction (RE) models for GDA extraction&lt;br /&gt;
are hand-labeled corpora [11, 12, 13]. However, hand-labeling data is an expensive process&lt;br /&gt;
requiring large amounts of time to expert biologists and, therefore, all of these datasets are&lt;br /&gt;
limited in size. To address this limitation, distant supervision has been proposed [14]. Under the&lt;br /&gt;
distant supervision paradigm, all the sentences mentioning the same pair of entities are labeled&lt;br /&gt;
by the corresponding relation stored within a source database. The assumption is that if two&lt;br /&gt;
entities participate in a relation, at least one sentence mentioning them conveys that relation.&lt;br /&gt;
As a consequence, distant supervision generates a large number of false positives, since not&lt;br /&gt;
all sentences express the relation between the considered entities. To counter false positives,&lt;br /&gt;
the RE task under distant supervision can be modeled as a Multi-Instance Learning (MIL)&lt;br /&gt;
problem [15, 16, 17, 18]. With MIL, the sentences containing two entities connected by a given&lt;br /&gt;
relation are collected into bags labeled with such relation. Grouping sentences into bags reduces&lt;br /&gt;
noise, as a bag of sentences is more likely to express a relation than a single sentence. Thus,&lt;br /&gt;
distant supervision alleviates manual annotation efforts, and MIL increases the robustness of&lt;br /&gt;
RE models to noise.&lt;br /&gt;
   Since the advent of distant supervision, several datasets for RE have been developed under&lt;br /&gt;
this paradigm for news and biomedical science domains [14, 19, 6]. Among biomedical ones,&lt;br /&gt;
the most relevant datasets are BioRel [19], a large-scale dataset for domain-general Biomedical&lt;br /&gt;
Relation Extraction (BioRE), and DTI [6], a large-scale dataset developed to extract Drug-Target&lt;br /&gt;
Interactions (DTIs). In the wake of such efforts, we created TBGA: a novel large-scale, semi-&lt;br /&gt;
automatically annotated dataset for GDA extraction based on DisGeNET. We chose DisGeNET as&lt;br /&gt;
source database since it is one of the most comprehensive databases for GDAs [20], integrating&lt;br /&gt;
several expert-curated resources.&lt;br /&gt;
   Then, we trained and tested several state-of-the-art RE models on TBGA to create a large&lt;br /&gt;
and realistic benchmark for GDA extraction. We built models using OpenNRE [21], an open&lt;br /&gt;
and extensible toolkit for Neural Relation Extraction (NRE). The choice of OpenNRE eases the&lt;br /&gt;
re-use of the dataset and the models developed for this work to future researchers. Finally, we&lt;br /&gt;
publicly released TBGA on Zenodo,1 whereas we stored source code and scripts to train and&lt;br /&gt;
test RE models in a publicly available GitHub repository.2&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
2. Dataset&lt;br /&gt;
TBGA is the first large-scale, semi-automatically annotated dataset for GDA extraction. The&lt;br /&gt;
dataset consists of three text files, corresponding to train, validation, and test sets, plus an&lt;br /&gt;
additional JSON file containing the mapping between relation names and IDs. Each record in&lt;br /&gt;
train, validation, or test files corresponds to a single GDA extracted from a sentence, and it is&lt;br /&gt;
represented as a JSON object with the following attributes:&lt;br /&gt;
&lt;br /&gt;
    • text : sentence from which the GDA was extracted.&lt;br /&gt;
&lt;br /&gt;
   1&lt;br /&gt;
       https://doi.org/10.5281/zenodo.5911097&lt;br /&gt;
   2&lt;br /&gt;
       https://github.com/GDAMining/gda-extraction/&lt;br /&gt;
�    • relation : relation name associated with the given GDA.&lt;br /&gt;
    • h : JSON object representing the gene entity, composed of:&lt;br /&gt;
          ∘ id : NCBI Entrez ID associated with the gene entity.&lt;br /&gt;
          ∘ name : NCBI official gene symbol associated with the gene entity.&lt;br /&gt;
          ∘ pos : list consisting of starting position and length of the gene mention within text.&lt;br /&gt;
    • t : JSON object representing the disease entity, composed of:&lt;br /&gt;
          ∘ id : UMLS Concept Unique Identifier (CUI) associated with the disease entity.&lt;br /&gt;
          ∘ name : UMLS preferred term associated with the disease entity.&lt;br /&gt;
          ∘ pos : list consisting of starting position and length of the disease mention within&lt;br /&gt;
            text.&lt;br /&gt;
&lt;br /&gt;
If a sentence contains multiple gene-disease pairs, the corresponding GDAs are split into separate&lt;br /&gt;
data records.&lt;br /&gt;
    Overall, TBGA contains over 200,000 instances and 100,000 bags. Table 1 reports per-relation&lt;br /&gt;
statistics for the dataset. Notice the large number of Not Associated (NA) instances. Regarding&lt;br /&gt;
gene and disease statistics, the most frequent genes are tumor suppressor genes, such as TP53&lt;br /&gt;
and CDKN2A, and (proto-)oncogenes, like EGFR and BRAF. Among the most frequent diseases,&lt;br /&gt;
we have neoplasms such as breast carcinoma, lung adenocarcinoma, and prostate carcinoma.&lt;br /&gt;
As a consequence, the most frequent GDAs are gene-cancer associations.&lt;br /&gt;
&lt;br /&gt;
Table 1&lt;br /&gt;
Per-relation statistics for TBGA. Statistics are reported separately for each data split.&lt;br /&gt;
      Granularity       Split         Therapeutic    Biomarker     Genomic Alterations         NA&lt;br /&gt;
                        Train                3,139        20,145                  32,831    122,149&lt;br /&gt;
      Sentence-level    Validation             402         2,279                   2,306     15,206&lt;br /&gt;
                        Test                   384         2,315                   2,209     15,608&lt;br /&gt;
                        Train                2,218        13,372                  12,759     56,698&lt;br /&gt;
      Bag-level         Validation             331         2,019                   1,147      6,994&lt;br /&gt;
                        Test                   308         2,068                   1,122      6,996&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
3. Experimental Setup&lt;br /&gt;
3.1. Benchmarks&lt;br /&gt;
We performed experiments on three different datasets: TBGA, DTI, and BioRel. We used TBGA&lt;br /&gt;
as a benchmark to evaluate RE models for GDA extraction under the MIL setting. On the other&lt;br /&gt;
hand, we used DTI and BioRel only to validate the soundness of our implementation of the&lt;br /&gt;
baseline models.&lt;br /&gt;
�3.2. Evaluation Measures&lt;br /&gt;
We evaluated RE models using the Area Under the Precision-Recall Curve (AUPRC). AUPRC&lt;br /&gt;
is a popular measure to evaluate distantly-supervised RE models, which has been adopted by&lt;br /&gt;
OpenNRE [21] and used in several works, such as [6, 19]. For experiments on TBGA, we also&lt;br /&gt;
computed Precision at k items (P@k).&lt;br /&gt;
&lt;br /&gt;
3.3. Aggregation Strategies&lt;br /&gt;
We adopted two different sentence aggregation strategies to use RE models under the MIL&lt;br /&gt;
setting: average-based (AVE) and attention-based (ATT) [22]. The average-based aggregation&lt;br /&gt;
assumes that all sentences within the same bag contribute equally to the bag-level representation.&lt;br /&gt;
In other words, the bag representation is the average of all its sentence representations. On&lt;br /&gt;
the other hand, the attention-based aggregation represents each bag as a weighted sum of&lt;br /&gt;
its sentence representations, where the attention weights are dynamically adjusted for each&lt;br /&gt;
sentence.&lt;br /&gt;
&lt;br /&gt;
3.4. Relation Extraction Models&lt;br /&gt;
We considered the main state-of-the-art RE models to perform experiments: CNN [23], PCNN [24],&lt;br /&gt;
BiGRU [25, 19, 6], BiGRU-ATT [26, 6], and BERE [6]. All models use pre-trained word embed-&lt;br /&gt;
dings to initialize word representations. On the other hand, Position Features (PFs), Position&lt;br /&gt;
Indicators (PIs), and unknown words are initialized using the normal distribution, whereas&lt;br /&gt;
blank words are initialized with zeros.&lt;br /&gt;
   We adopted pre-trained BioWordVec [27] embeddings to perform experiments on TBGA. Two&lt;br /&gt;
versions of pre-trained BioWordVec embeddings are available: “Bio_embedding_intrinsic” and&lt;br /&gt;
“Bio_embedding_extrinsic”. We chose the “Bio_embedding_extrinsic” version as it is the most&lt;br /&gt;
suitable for BioRE. As for the experiments on DTI and BioRel, we adopted the pre-trained word&lt;br /&gt;
embeddings used in the original works [6, 19] – that is, the word embeddings from Pyysalo et&lt;br /&gt;
al. [28] for DTI, and the “Bio_embedding_extrinsic” version of BioWordVec for BioRel.&lt;br /&gt;
   For TBGA experiments, we used grid search to determine the best combination between&lt;br /&gt;
optimizer and learning rate. As combinations, we tested Stochastic Gradient Descent (SGD)&lt;br /&gt;
with learning rate among {0.1, 0.2, 0.3, 0.4, 0.5} and Adam [29] with learning rate set to 0.0001.&lt;br /&gt;
For all RE models, we set the rest of the hyper-parameters empirically.&lt;br /&gt;
   For DTI and BioRel experiments, we relied on the hyper-parameter settings reported in the&lt;br /&gt;
original works [6, 19].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
4. Experimental Results&lt;br /&gt;
We report the results for two different experiments. The first experiment aims to validate the&lt;br /&gt;
soundness of the implementation of the considered RE models. To this end, we trained and&lt;br /&gt;
tested the RE models on DTI and BioRel datasets, and we compared the AUPRC scores we&lt;br /&gt;
obtained against those reported in the original works [6, 19]. For this experiment, we only&lt;br /&gt;
compared the RE models and aggregation strategies that were used in the original works. The&lt;br /&gt;
�Table 2&lt;br /&gt;
Results of the baselines validation on DTI [6] and BioRel [19] datasets. The “–” symbol means that the&lt;br /&gt;
RE model, for the given aggregation strategy, has not been originally evaluated on the specific dataset.&lt;br /&gt;
                     Model          Strategy   Implementation       DTI    BioRel&lt;br /&gt;
                                               Reproduced             –     0.800&lt;br /&gt;
                                    AVE&lt;br /&gt;
                                               Original               –     0.790&lt;br /&gt;
                     CNN&lt;br /&gt;
                                               Reproduced             –     0.790&lt;br /&gt;
                                    ATT&lt;br /&gt;
                                               Original               –     0.780&lt;br /&gt;
                                               Reproduced          0.234    0.860&lt;br /&gt;
                                    AVE&lt;br /&gt;
                                               Original            0.160    0.820&lt;br /&gt;
                     PCNN&lt;br /&gt;
                                               Reproduced          0.408    0.820&lt;br /&gt;
                                    ATT&lt;br /&gt;
                                               Original            0.359    0.790&lt;br /&gt;
                                               Reproduced              –    0.870&lt;br /&gt;
                                    AVE&lt;br /&gt;
                                               Original                –    0.800&lt;br /&gt;
                     BiGRU&lt;br /&gt;
                                               Reproduced          0.379    0.850&lt;br /&gt;
                                    ATT&lt;br /&gt;
                                               Original            0.390    0.780&lt;br /&gt;
                                               Reproduced          0.383        –&lt;br /&gt;
                     BiGRU-ATT      ATT&lt;br /&gt;
                                               Original            0.457        –&lt;br /&gt;
                                               Reproduced          0.407        –&lt;br /&gt;
                                    AVE&lt;br /&gt;
                                               Original            0.384        –&lt;br /&gt;
                     BERE&lt;br /&gt;
                                               Reproduced          0.525        –&lt;br /&gt;
                                    ATT&lt;br /&gt;
                                               Original            0.524        –&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
second experiment uses TBGA as a benchmark to evaluate RE models for GDA extraction. In&lt;br /&gt;
this case, we trained and tested all the considered RE models using both aggregation strategies.&lt;br /&gt;
For each RE model, we reported the AUPRC and P@k scores.&lt;br /&gt;
&lt;br /&gt;
4.1. Baselines Validation&lt;br /&gt;
The results of the baselines validation are reported in Table 2. We can observe that the RE&lt;br /&gt;
models we use from – or implement within – OpenNRE achieve performance higher than or&lt;br /&gt;
comparable to those reported in DTI and BioRel original works. The only exceptions are BiGRU&lt;br /&gt;
and BiGRU-ATT on DTI, where the AUPRC scores of our implementations are lower than those&lt;br /&gt;
reported in the original work. However, Hong et al. [6] report the optimal hyper-parameter&lt;br /&gt;
settings for BERE, but not for the baselines. Thus, we attribute the negative difference between&lt;br /&gt;
our implementations and theirs to the lack of information about optimal hyper-parameters.&lt;br /&gt;
Overall, the results confirm the soundness of our implementations. Therefore, we can consider&lt;br /&gt;
them as competitive baseline models to use for benchmarking GDA extraction.&lt;br /&gt;
&lt;br /&gt;
4.2. GDA Benchmarking&lt;br /&gt;
Table 3 reports the AUPRC and P@k scores of RE models on TBGA. Given the RE models&lt;br /&gt;
performance, we make the following observations. First, the AUPRC performances achieved by&lt;br /&gt;
�Table 3&lt;br /&gt;
RE models performance on TBGA dataset. For each measure, bold values represent the best scores.&lt;br /&gt;
        Model         Strategy   AUPRC      P@50    P@100    P@250     P@500    P@1000&lt;br /&gt;
                      AVE           0.422   0.780    0.760     0.744    0.696      0.625&lt;br /&gt;
        CNN&lt;br /&gt;
                      ATT           0.403   0.780    0.760     0.788    0.710      0.624&lt;br /&gt;
                      AVE           0.426   0.780    0.780     0.744    0.720      0.664&lt;br /&gt;
        PCNN&lt;br /&gt;
                      ATT           0.404   0.760    0.750     0.744    0.700      0.628&lt;br /&gt;
                      AVE           0.437   0.620    0.720     0.724    0.730      0.678&lt;br /&gt;
        BiGRU&lt;br /&gt;
                      ATT           0.423   0.760    0.750     0.748    0.726      0.666&lt;br /&gt;
                      AVE           0.419   0.740    0.740     0.748    0.694      0.615&lt;br /&gt;
        BiGRU-ATT&lt;br /&gt;
                      ATT           0.390   0.680    0.760     0.756    0.702      0.631&lt;br /&gt;
                      AVE          0.419    0.700    0.710    0.720     0.704     0.620&lt;br /&gt;
        BERE&lt;br /&gt;
                      ATT          0.445    0.780    0.780    0.800     0.764     0.709&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
RE models on TBGA indicate a high complexity of the GDA extraction task. The task complexity&lt;br /&gt;
is further supported by the lower performances obtained by top-performing RE models on&lt;br /&gt;
TBGA compared to DTI and BioRel (cf. Table 2). Secondly, CNN, PCNN, BiGRU, and BiGRU-&lt;br /&gt;
ATT RE models behave similarly. Among them, BiGRU-ATT has the worst performance. This&lt;br /&gt;
suggests that replacing BiGRU max pooling layer with an attention layer proves less effective.&lt;br /&gt;
Overall, the best AUPRC and P@k scores are achieved by BERE when using the attention-&lt;br /&gt;
based aggregation strategy. This highlights BERE effectiveness of fully exploiting sentence&lt;br /&gt;
information from both semantic and syntactic aspects [6]. Thirdly, in terms of AUPRC, the&lt;br /&gt;
attention-based aggregation proves less effective than the average-based one. On the other hand,&lt;br /&gt;
attention-based aggregation provides mixed results on P@k measures. Although in contrast&lt;br /&gt;
with the results obtained in general-domain RE [22], this trend is in line with the results found&lt;br /&gt;
by Xing et al. [19] on BioRel, where RE models using an average-based aggregation strategy&lt;br /&gt;
achieve performance comparable to or higher than those using an attention-based one. The&lt;br /&gt;
only exception is BERE, whose performance using the attention-based aggregation outperforms&lt;br /&gt;
the one using the average-based strategy. Thus, the obtained results suggest that TBGA is a&lt;br /&gt;
challenging dataset for GDA extraction.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
5. Conclusions&lt;br /&gt;
We have created TBGA, a large-scale, semi-automatically annotated dataset for GDA extraction.&lt;br /&gt;
Automatic GDA extraction is one of the most relevant tasks of BioRE. We have used TBGA as a&lt;br /&gt;
benchmark to evaluate state-of-the-art BioRE models on GDA extraction. The results suggest&lt;br /&gt;
that TBGA is a challenging dataset for this task and, in general, for BioRE.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Acknowledgments&lt;br /&gt;
The work was supported by the EU H2020 ExaMode project, under Grant Agreement no. 825292.&lt;br /&gt;
�References&lt;br /&gt;
 [1] S. Marchesin, G. Silvello, TBGA: a large-scale gene-disease association dataset for biomed-&lt;br /&gt;
     ical relation extraction, BMC Bioinform. 23 (2022) 111.&lt;br /&gt;
 [2] A. Bairoch, R. Apweiler, The SWISS-PROT protein sequence data bank and its supplement&lt;br /&gt;
     TrEMBL, Nucleic Acids Res. 25 (1997) 31–36.&lt;br /&gt;
 [3] D. S. Wishart, C. Knox, A. Guo, S. Shrivastava, M. Hassanali, P. Stothard, Z. Chang,&lt;br /&gt;
     J. Woolsey, DrugBank: a comprehensive resource for in silico drug discovery and explo-&lt;br /&gt;
     ration, Nucleic Acids Res. 34 (2006) 668–672.&lt;br /&gt;
 [4] C. J. Mattingly, G. T. Colby, J. N. Forrest, J. L. Boyer, The Comparative Toxicogenomics&lt;br /&gt;
     Database (CTD), Environ. Health Perspect. 111 (2003) 793–795.&lt;br /&gt;
 [5] P. Buneman, J. Cheney, W. C. Tan, S. Vansummeren, Curated Databases, in: Proc. of the&lt;br /&gt;
     Twenty-Seventh ACM SIGMOD-SIGACT-SIGART Symposium on Principles of Database&lt;br /&gt;
     Systems, PODS 2008, June 9-11, 2008, Vancouver, BC, Canada, ACM, 2008, pp. 1–12.&lt;br /&gt;
 [6] L. Hong, J. Lin, S. Li, F. Wan, H. Yang, T. Jiang, D. Zhao, J. Zeng, A novel machine learn-&lt;br /&gt;
     ing framework for automated biomedical relation extraction from large-scale literature&lt;br /&gt;
     repositories, Nat. Mach. Intell. 2 (2020) 347–355.&lt;br /&gt;
 [7] E. P. Barracchia, G. Pio, D. D’Elia, M. Ceci, Prediction of new associations between ncrnas&lt;br /&gt;
     and diseases exploiting multi-type hierarchical clustering, BMC Bioinform. 21 (2020) 70.&lt;br /&gt;
 [8] S. Alaimo, R. Giugno, A. Pulvirenti, ncpred: ncrna-disease association prediction through&lt;br /&gt;
     tripartite network-based inference, Front. Bioeng. Biotechnol. 2 (2014).&lt;br /&gt;
 [9] S. Dugger, A. Platt, D. Goldstein, Drug development in the era of precision medicine, Nat.&lt;br /&gt;
     Rev. Drug. Discov. 17 (2018) 183–196.&lt;br /&gt;
[10] J. P. González, J. M. Ramírez-Anguita, J. Saüch-Pitarch, F. Ronzano, E. Centeno, F. Sanz, L. I.&lt;br /&gt;
     Furlong, The DisGeNET knowledge platform for disease genomics: 2019 update, Nucleic&lt;br /&gt;
     Acids Res. 48 (2020) D845–D855.&lt;br /&gt;
[11] E. M. van Mulligen, A. Fourrier-Réglat, D. Gurwitz, M. Molokhia, A. Nieto, G. Trifirò, J. A.&lt;br /&gt;
     Kors, L. I. Furlong, The EU-ADR corpus: Annotated drugs, diseases, targets, and their&lt;br /&gt;
     relationships, J. Biomed. Informatics 45 (2012) 879–884.&lt;br /&gt;
[12] D. Cheng, C. Knox, N. Young, P. Stothard, S. Damaraju, D. S. Wishart, PolySearch: a&lt;br /&gt;
     web-based text mining system for extracting relationships between human diseases, genes,&lt;br /&gt;
     mutations, drugs and metabolites, Nucleic Acids Res. 36 (2008) 399–405.&lt;br /&gt;
[13] H. J. Lee, S. H. Shim, M. R. Song, H. Lee, J. C. Park, CoMAGC: a corpus with multi-faceted&lt;br /&gt;
     annotations of gene-cancer relations, BMC Bioinform. 14 (2013) 323.&lt;br /&gt;
[14] M. Mintz, S. Bills, R. Snow, D. Jurafsky, Distant supervision for relation extraction without&lt;br /&gt;
     labeled data, in: Proc. of the 47th Annual Meeting of the Association for Computational&lt;br /&gt;
     Linguistics (ACL 2009) and the 4th International Joint Conference on Natural Language&lt;br /&gt;
     Processing of the AFNLP, 2-7 August 2009, Singapore, ACL, 2009, pp. 1003–1011.&lt;br /&gt;
[15] T. G. Dietterich, R. H. Lathrop, T. Lozano-Pérez, Solving the Multiple Instance Problem&lt;br /&gt;
     with Axis-Parallel Rectangles, Artif. Intell. 89 (1997) 31–71.&lt;br /&gt;
[16] S. Riedel, L. Yao, A. McCallum, Modeling Relations and Their Mentions without Labeled&lt;br /&gt;
     Text, in: Proc. of Machine Learning and Knowledge Discovery in Databases, European&lt;br /&gt;
     Conference, ECML PKDD 2010, Barcelona, Spain, September 20-24, 2010, volume 6323 of&lt;br /&gt;
     LNCS, Springer, 2010, pp. 148–163.&lt;br /&gt;
�[17] R. Hoffmann, C. Zhang, X. Ling, L. S. Zettlemoyer, D. S. Weld, Knowledge-Based Weak&lt;br /&gt;
     Supervision for Information Extraction of Overlapping Relations, in: The 49th Annual&lt;br /&gt;
     Meeting of the Association for Computational Linguistics: Human Language Technologies,&lt;br /&gt;
     Proceedings of the Conference, 19-24 June, 2011, Portland, Oregon, USA, ACL, 2011, pp.&lt;br /&gt;
     541–550.&lt;br /&gt;
[18] M. Surdeanu, J. Tibshirani, R. Nallapati, C. D. Manning, Multi-instance Multi-label Learning&lt;br /&gt;
     for Relation Extraction, in: Proc. of the 2012 Joint Conference on Empirical Methods in&lt;br /&gt;
     Natural Language Processing and Computational Natural Language Learning, EMNLP-&lt;br /&gt;
     CoNLL 2012, July 12-14, 2012, Jeju Island, Korea, ACL, 2012, pp. 455–465.&lt;br /&gt;
[19] R. Xing, J. Luo, T. Song, BioRel: towards large-scale biomedical relation extraction, BMC&lt;br /&gt;
     Bioinform. 21-S (2020) 543.&lt;br /&gt;
[20] Z. Tanoli, U. Seemab, A. Scherer, K. Wennerberg, J. Tang, M. Vähä-Koskela, Exploration of&lt;br /&gt;
     databases and methods supporting drug repurposing: a comprehensive survey, Briefings&lt;br /&gt;
     Bioinform. 22 (2021) 1656–1678.&lt;br /&gt;
[21] X. Han, T. Gao, Y. Yao, D. Ye, Z. Liu, M. Sun, OpenNRE: An Open and Extensible Toolkit&lt;br /&gt;
     for Neural Relation Extraction, in: Proc. of the 2019 Conference on Empirical Methods&lt;br /&gt;
     in Natural Language Processing and the 9th International Joint Conference on Natural&lt;br /&gt;
     Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019, ACL,&lt;br /&gt;
     2019, pp. 169–174.&lt;br /&gt;
[22] Y. Lin, S. Shen, Z. Liu, H. Luan, M. Sun, Neural Relation Extraction with Selective Attention&lt;br /&gt;
     over Instances, in: Proc. of the 54th Annual Meeting of the Association for Computational&lt;br /&gt;
     Linguistics, ACL 2016, August 7-12, 2016, Berlin, Germany, Volume 1: Long Papers, ACL,&lt;br /&gt;
     2016, pp. 2124–2133.&lt;br /&gt;
[23] D. Zeng, K. Liu, S. Lai, G. Zhou, J. Zhao, Relation Classification via Convolutional Deep&lt;br /&gt;
     Neural Network, in: Proc. of COLING 2014, 25th International Conference on Computa-&lt;br /&gt;
     tional Linguistics, Technical Papers, August 23-29, 2014, Dublin, Ireland, ACL, 2014, pp.&lt;br /&gt;
     2335–2344.&lt;br /&gt;
[24] D. Zeng, K. Liu, Y. Chen, J. Zhao, Distant Supervision for Relation Extraction via Piecewise&lt;br /&gt;
     Convolutional Neural Networks, in: Proc. of the 2015 Conference on Empirical Methods&lt;br /&gt;
     in Natural Language Processing, EMNLP 2015, Lisbon, Portugal, September 17-21, 2015,&lt;br /&gt;
     ACL, 2015, pp. 1753–1762.&lt;br /&gt;
[25] D. Zhang, D. Wang, Relation Classification via Recurrent Neural Network, CoRR&lt;br /&gt;
     abs/1508.01006 (2015).&lt;br /&gt;
[26] P. Zhou, W. Shi, J. Tian, Z. Qi, B. Li, H. Hao, B. Xu, Attention-Based Bidirectional Long&lt;br /&gt;
     Short-Term Memory Networks for Relation Classification, in: Proc. of the 54th Annual&lt;br /&gt;
     Meeting of the Association for Computational Linguistics, ACL 2016, August 7-12, 2016,&lt;br /&gt;
     Berlin, Germany, Volume 2: Short Papers, ACL, 2016, pp. 207–212.&lt;br /&gt;
[27] Y. Zhang, Q. Chen, Z. Yang, H. Lin, Z. Lu, BioWordVec, improving biomedical word&lt;br /&gt;
     embeddings with subword information and MeSH, Sci Data 6 (2019) 1–9.&lt;br /&gt;
[28] S. Pyysalo, F. Ginter, H. Moen, T. Salakoski, S. Ananiadou, Distributional Semantics&lt;br /&gt;
     Resources for Biomedical Text Processing, Proc. of LBM (2013) 39–44.&lt;br /&gt;
[29] D. P. Kingma, J. Ba, Adam: A Method for Stochastic Optimization, in: Proc. of the 3rd&lt;br /&gt;
     International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA,&lt;br /&gt;
     May 7-9, 2015, 2015, pp. 1–15.&lt;br /&gt;
�&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;/div&gt;</summary>
		<author><name>Wf</name></author>
	</entry>
</feed>