Difference between revisions of "ProceedingsTitleParser"
Line 324: | Line 324: | ||
== Data Analysis == | == Data Analysis == | ||
=== Year === | === Year === | ||
− | <source lang='sql' highlight="1- | + | <source lang='sql' highlight="1-3,5-10" lines> |
select year,count(*) from event | select year,count(*) from event | ||
group by year | group by year | ||
Line 445: | Line 445: | ||
(null) 24389 | (null) 24389 | ||
</source> | </source> | ||
+ | |||
== Choice of Database/Storage system == | == Choice of Database/Storage system == | ||
The following candidates for a Database/storage system where considered: | The following candidates for a Database/storage system where considered: |
Revision as of 14:06, 20 October 2020
OsProject | |
---|---|
edit | |
id | ProceedingsTitleParser |
state | |
owner | WolfgangFahl |
title | Shallow Semantic Parser to extract metadata from scientific proceedings titles |
url | https://github.com/WolfgangFahl/ProceedingsTitleParser |
version | 0.0.1 |
description | |
date | 2020-07-02 |
since | |
until |
Usage
What is it?
The Proceedings Title Parser Service is a specialized search engine for scientific proceedings and events. It searches in a corpus/database based on data sourced from
- http://www.openresearch.org ✓
- http://ceur-ws.org ✓
- http://www.wikidata.org (in progress)
- http://confref.org ✓
- http://crossref.org ✓
- https://dblp.org/ (in progress)
- GND (in progress)
- http://www.wikicfp.com/cfp/ ✓
- ...
Search Modes
The Proceedings Title Parser currently has three modes:
- Proceedings Title Parsing
- Named Entity Recognition (NER)
- Extract/Scrape
All three modes expect some lines of text as input. The mode is automatically detected/selected by the content of the lines given as input.
Example
The input:
Proceedings of the 2020 International Conference on High Performance Computing & Simulation (HPCS 2020) BIR 2019 http://ceur-ws.org/Vol-2599/
- will trigger Proceedings Title Parsing mode for the first line
- Named Entity Recognition mode for the second line
- Extract/Scape mode for the third line
Proceedings Title Parsing mode
In Proceedings Title Parsing mode the content of a line is parsed word by word to find typical elements of Proceedings titles like
- ordinal (First, III., 6th, 23rd)
- city (Paris, New York, London, Berlin, Barcelona)
- country (USA, Italy, Germany, France, UK)
- provinces (California, Texas, Florida, Ontario, NRW)
- scope (International, European, Czech, Italian, National)
- year (2016,1988,2017,2018,2019,2020)
- ...
while the above elements can be found by comparing with a list of known cities, countries, provinces, ... finding the event acronym needs a lookup in a corpus/database of proceedings and events which is automatically performed to check whether a known event acronym like ISWC, ICEIS, SIGSPATIAL, ... might be found with the given context e.g. year. If the acronym (with year) is found a link to the resulting source record is shown.
Example
Input
Proceedings of the 2020 International Conference on High Performance Computing & Simulation (HPCS 2020)
# | Source | Acronym | Url | Found by |
---|---|---|---|---|
1 | OPEN RESEARCH | HPCS 2020 | https://www.openresearch.org/wiki/HPCS%202020 | HPCS 2020 |
{'year': '2020', 'scope': 'International', 'event': 'Conference', 'topic': 'High Performance Computing & Simulation', 'acronym': 'HPCS 2020', 'title': 'Proceedings of the 2020 International Conference on High Performance Computing & Simulation (HPCS 2020)', 'source': 'line', 'publish': 'Proceedings', 'syntax': 'on', 'delimiter': '&'}
{ "acronym": "HPCS 2020", "city": "Barcelona", "country": "Spain", "creation_date": "2020-03-26 06:01:12", "end_date": "2020-07-24 00:00:00", "event": "HPCS 2020", "foundBy": "HPCS 2020", "homePage": null, "homepage": "http://conf.cisedu.info/rp/hpcs20/", "modification_date": "2020-03-26 08:16:33", "series": "HPCS", "source": "OPEN RESEARCH", "start_date": "2020-07-20 00:00:00", "title": "2020 International Conference on High Performance Computing & Simulation", "url": "https://www.openresearch.org/wiki/HPCS 2020" }
Named Entity Recognition mode (NER)
In named entity recognition mode the words to be looked up can be directly entered without following the patterns of typical proceedings titles. Syntax elements like "Proceedings of ... " may be left out. This mode will often also give good results but can not use the information provided by syntactically elements like "at". For an example "Proceedings of the 1st conference of the history of Rome at Paris" the NER equivalent "1 conference history Rome Paris" will obviously be ambiguous.
Example
Input
BIR 2019
Result
# | Source | Acronym | Url | Found by |
---|---|---|---|---|
1 | OPEN RESEARCH | BIR 2019 | https://www.openresearch.org/wiki/BIR%202019 | BIR 2019 |
2 | BIR 2019 | BIR 2019 | http://ceur-ws.org/Vol-2345 | BIR 2019 |
3 | confref | BIR 2019 | http://portal.confref.org/list/bir2019 | BIR 2019 |
'title': 'BIR 2019', 'source': 'line', 'year': '2019'}
{ "acronym": "BIR 2019", "city": null, "country": null, "creation_date": "2020-03-09 11:01:20", "end_date": "2019-09-25 00:00:00", "event": "BIR 2019", "foundBy": "BIR 2019", "homepage": "https://bir2019.ue.katowice.pl/", "modification_date": "2020-07-06 11:47:07", "series": "BIR", "source": "OPEN RESEARCH", "start_date": "2019-09-23 00:00:00", "title": "18th International Conference on Perspectives in Business Informatics Research", "url": "https://www.openresearch.org/wiki/BIR 2019" }
{ "acronym": "BIR 2019", "city": "Cologne", "country": "Germany", "enum": "8th", "event": "BIR 2019", "eventId": "Vol-2345", "eventType": "Workshop", "foundBy": "BIR 2019", "homePage": null, "month": "April", "ordinal": 14, "publish": "Proceedings", "scope": "International", "source": "CEUR-WS", "syntax": "on", "title": "Proceedings of the 8th International Workshop on Bibliometric-enhanced Information Retrieval (BIR 2019)co-located with the 41st European Conference on Information Retrieval (ECIR 2019),Cologne, Germany, April 14th, 2019.Submitted by: Guillaume Cabanac", "topic": "Bibliometric-enhanced Information Retrieval", "url": "http://ceur-ws.org/Vol-2345", "year": "2019" }
{ "acronym": "BIR 2019", "address": null, "area": { "id": 2, "value": "Computer Science" }, "cameraReadyDate": null, "city": "Katowice", "confSeries": { "dblpId": "https://dblp.org/db/conf/bir/", "description": null, "eissn": null, "id": "bir", "issn": null, "name": "Business Informatics Research" }, "country": "Poland", "description": null, "endDate": "2019-09-25", "event": "BIR 2019", "foundBy": "BIR 2019", "homepage": null, "id": "bir2019", "keywords": [ "Data mining", "Enterprise architecture", "Business process model", "Aggregation", "Acceptance", "Banking", "Blockchain", "Comparison", "Digital learning", "e-Health", "Agile modelling method engineering", "Literature review", "Digitalization", "ecosystem", "Barriers of change", "Bing", "Business process management system", "digital Workplace Health Promotion (dWHP)", "Dual video cast", "e-Lecture" ], "name": "Business Informatics Research", "notificationDate": null, "ranks": [], "shortDescription": null, "source": "confref", "startDate": "2019-09-23", "submissionDate": null, "submissionExtended": false, "url": "http://portal.confref.org/list/bir2019", "year": 2019 }
Extract / scrape mode
If the line contains an url of a known source of conference of proceedings title information the page will be automatically visited and the meta data extracted.
Example
Input
http://ceur-ws.org/Vol-2599/
Result:
# | Source | Acronym | Url | Found by |
---|---|---|---|---|
1 | CEUR-WS | BlockSW | http://ceur-ws.org/Vol-2599 | BlockSW |
{'prefix': 'Blockchain enabled Semantic Web', 'event': 'Workshop', 'acronym': 'BlockSW', 'title': 'Proceedings of the Blockchain enabled Semantic Web Workshop (BlockSW) and Contextualized Knowledge Graphs (CKG) Workshop', 'source': 'CEUR-WS', 'eventId': 'Vol-2599', 'publish': 'Proceedings', 'syntax': 'and'}
{ "acronym": "BlockSW", "event": "BlockSW", "eventId": "Vol-2599", "eventType": "Workshop", "foundBy": "BlockSW", "homePage": null, "month": "October", "prefix": "Blockchain enabled Semantic Web", "publish": "Proceedings", "source": "CEUR-WS", "syntax": "New", "title": "Proceedings of the Blockchain enabled Semantic Web Workshop (BlockSW) and Contextualized Knowledge Graphs (CKG) Workshop (BlockSW-CKG 2019),Auckland, New Zealand, October 27, 2019.Submitted by: Reza Samavi", "url": "http://ceur-ws.org/Vol-2599", "year": "2019" }
Result formats / Content Negotiation
The following result formats are supported:
- html (default)
- csv
- json
- xml
To select a result format you can either add the "&format" query parameter as part of your url or specify the corresponding accept header in your query.
Example format parameter queries
csv
http://ptp.bitplan.com/parse?examples=example2&titles=PAKM+2000&format=csv
result is the same as with content-negotiation text/csv
json
http://ptp.bitplan.com/parse?examples=example2&titles=BIR+2019&format=json
Try it! result is the same as with content-negotiation application/json
xml
http://ptp.bitplan.com/parse?examples=example2&titles=EuroPar+2020&format=xml
Try it! result is the same as with content-negotiation application/xml or text/xml
wikison
To get WikiSon format
http://ptp.bitplan.com/parse?examples=example2&titles=EuroPar+2020&format=wikison
Result:
{{Event
|homepage=https://2020.euro-par.org/
|event=EuroPar 2020
|series=EuroPar
|acronym=EuroPar 2020
|title=International European Conference on Parallel and Distributed Computing
|city=Warsaw
|country=Poland
|start_date=2020-08-24 00:00:00
|end_date=2020-08-28 00:00:00
|url=https://www.openresearch.org/wiki/EuroPar 2020
}}
Examples with accept header
csv
curl -H "Accept: text/csv" "http://ptp.bitplan.com/parse?examples=example2&titles=PAKM+2000"
Result for text/csv
"month","homePage","eventType","country","acronym","ordinal","url","year","event","eventId","source","syntax","enum","foundBy","location","publish","scope","title" "October","","Conference","Switzerland","PAKM 2000",3,"http://ceur-ws.org/Vol-34","2000","PAKM 2000","Vol-34","CEUR-WS","the","Third","PAKM 2000","Basel","Proceedings","International","Proceedings of the Third International Conference (PAKM 2000), Basel, Switzerland, October 30-31, 2000.Submitted by: Ulrich Reimer"
json
curl -H "Accept: application/json" "http://ptp.bitplan.com/parse?examples=example2&titles=BIR+2019"
Result for application/json
{
"count": 3,
"events": [{
"acronym": "BIR 2019",
"city": null,
"country": null,
"creation_date": "2020-03-09T11:01:20+00:00",
"end_date": "2019-09-25T00:00:00+00:00",
"event": "BIR 2019",
"foundBy": "BIR 2019",
"homePage": null,
"homepage": "https://bir2019.ue.katowice.pl/",
"modification_date": "2020-07-06T11:47:07+00:00",
"series": "BIR",
"source": "OPEN RESEARCH",
"start_date": "2019-09-23T00:00:00+00:00",
"title": "18th International Conference on Perspectives in Business Informatics Research",
"url": "https://www.openresearch.org/wiki/BIR 2019"
}, {
"acronym": "BIR 2019",
"city": "Cologne",
"country": "Germany",
"enum": "8th",
"event": "BIR 2019",
"eventId": "Vol-2345",
"eventType": "Workshop",
"foundBy": "BIR 2019",
"homePage": null,
"month": "April",
"ordinal": 14,
"publish": "Proceedings",
"scope": "International",
"source": "CEUR-WS",
"syntax": "on",
"title": "Proceedings of the 8th International Workshop on Bibliometric-enhanced Information Retrieval (BIR 2019)co-located with the 41st European Conference on Information Retrieval (ECIR 2019),Cologne, Germany, April 14th, 2019.Submitted by: Guillaume Cabanac",
"topic": "Bibliometric-enhanced Information Retrieval",
"url": "http://ceur-ws.org/Vol-2345",
"year": "2019"
}, {
"acronym": "BIR 2019",
"address": null,
"area": {
"value": "Computer Science",
"id": 2
},
"cameraReadyDate": null,
"city": "Katowice",
"confSeries": {
"id": "bir",
"issn": null,
"eissn": null,
"dblpId": "https://dblp.org/db/conf/bir/",
"name": "Business Informatics Research",
"description": null
},
"country": "Poland",
"description": null,
"endDate": "2019-09-25",
"event": "BIR 2019",
"foundBy": "BIR 2019",
"homepage": null,
"id": "bir2019",
"keywords": ["Data mining", "Enterprise architecture", "Business process model", "Aggregation", "Acceptance", "Banking", "Blockchain", "Comparison", "Digital learning", "e-Health", "Agile modelling method engineering", "Literature review", "Digitalization", "ecosystem", "Barriers of change", "Bing", "Business process management system", "digital Workplace Health Promotion (dWHP)", "Dual video cast", "e-Lecture"],
"name": "Business Informatics Research",
"notificationDate": null,
"ranks": [],
"shortDescription": null,
"source": "confref",
"startDate": "2019-09-23",
"submissionDate": null,
"submissionExtended": false,
"url": "http://portal.confref.org/list/bir2019",
"year": 2019
}]
}
xml
As an alternative to application/xml the mime-type text/xml is also accepted with the same result.
curl -H "Accept: application/xml" "http://ptp.bitplan.com/parse?examples=example2&titles=EuroPar+2020"
Result for application/xml
<?xml version="1.0" ?>
<events>
<event>
<foundBy>EuroPar 2020</foundBy>
<homepage>https://2020.euro-par.org/</homepage>
<event>EuroPar 2020</event>
<series>EuroPar</series>
<acronym>EuroPar 2020</acronym>
<title>International European Conference on Parallel and Distributed Computing</title>
<city>Warsaw</city>
<country>Poland</country>
<start_date>2020-08-24T00:00:00</start_date>
<end_date>2020-08-28T00:00:00</end_date>
<creation_date>2020-02-27T14:44:52</creation_date>
<modification_date>2020-02-27T14:44:52</modification_date>
<url>https://www.openresearch.org/wiki/EuroPar 2020</url>
<source>OPEN RESEARCH</source>
</event>
</events>
Running your own service
PreRequisites
If you'd like to run your own copy of this service you'll need:
- git
- python 3 (>=3.6)
- some unix command line tools like curl, grep, wc
Tested on Linux (Ubuntu bionic/Travis) and MacOS High Sierra 10.13.6 (using macports)
Installation
git clone https://github.com/WolfgangFahl/ProceedingsTitleParser
./install
Getting the sample data
Getting the sample data from the different sources may take a few minutes. You'll need som 60 MBytes of disk space as of 2020-07-11
scripts/getsamples
Testing
./test
Implementation
Data Analysis
Year
select year,count(*) from event
group by year
order by 1 desc
year count(*)
19670 1
2109 1
2106 1
2105 1
2091 1
2088 1
2081 1
2026 3
2025 1
2024 2
2022 3
2021 1069
2020 8318
2019 19032
2018 19546
2017 17618
2016 15697
2015 14221
2014 13831
2013 12621
2012 12292
2011 11926
2010 10416
2009 9198
2008 9024
2007 5569
2006 4365
2005 3943
2004 3438
2003 3092
2002 2782
2001 2519
2000 2400
1999 2111
1998 2063
1997 1986
1996 1732
1995 1612
1994 1578
1993 1474
1992 1308
1991 1194
1990 1063
1989 1061
1988 1029
1987 704
1986 715
1985 608
1984 486
1983 460
1982 405
1981 346
1980 314
1979 302
1978 276
1977 197
1976 152
1975 144
1974 113
1973 102
1972 82
1971 78
1970 66
1969 66
1968 54
1967 55
1966 41
1965 32
1964 31
1963 29
1962 28
1961 19
1960 17
1959 10
1958 9
1957 8
1956 8
1955 7
1954 3
1953 5
1952 4
1951 5
1950 4
1949 2
1948 2
1947 1
1941 1
1938 1
1935 1
1932 1
1930 1
1929 1
1926 1
1923 1
1922 1
1920 2
1914 1
1913 1
1912 1
1911 1
1910 1
1908 1
1905 1
1904 1
1901 1
1900 146
1895 1
1894 1
1890 1
1889 1
1880 1
1862 1
35 1
0 14
(null) 24389
Choice of Database/Storage system
The following candidates for a Database/storage system where considered:
- Python native solutions
- using JSON files as caches
- ruruki
- Graph databases
- RDF TripleStore
- SQL Database
To investigate the feasibility of the approaches the Open Source Project DgraphAndWeaviateTest was created and issues for the different providers where created and questions on https://stackoverflow.com asked and answers given.
Evaluation results
Native JSON Files
Native Json Files have a very good performance and are simple to handle e.g. reading the crossRef data takes approx 1 sec
read 45601 events in 1.1 s
The disadvantage is that the query capabilities of the JSON only approach are limited. In the protype only lookup by eventID and eventAcronym was implemented via HashTables/Dicts.
Ruruki
Searching for an "in memory graph database solution for python" we found ruruki. The performance was disappointing. See testSpeed
creating 1000 vertices took 1.7 s = 598.3 v/s
Gremlin python
Using Gremlin would have been our favorite choice based on the experience with the SimpleGraph project. From our experience in-memory gremlin works excellent with less than 100.000 vertices when using Java. It is still useable for around 1 million vertices. If there are a lot more vertices it's necessay to back gremlin with a proper Graph database. We were not successful with any of the Graph database yet. The situation get's worse when Python is used as a programming language see: Gremlin python. Python is only supported as a language variant and it's very awkward to work in that environment. From our point of view this is not production ready so we didn't bother to investigate further.
DGraph
Dgraph looks like a good candidate for a graph database. It's simple to handle, has a docker based installation and a python interface. The tests looked very promising but in our development environment there has been some instability as reported in unreliability issue for Dgraph. It's not clear what causes this problem but it shows up in an undeterministic way in the travis CI tests and is a showstopper at this time. We will certainly reconsider Dgraph later.
Weaviate
Weaviate looks like an excellent fit for our usecase since it has NLP support built in and comes with a "contextionary" for e.g. english out of the box. That means the dictionary approach of the Proceedings Title Parser would have GloVe support immediately.
Apache Jena
Apache Jena has an RDF triplestore with SPARQL query capabilities. The question was whether the integration with Python and the "List of Dicts" based approach of the EventManager would be feasible. The github issues
- Add Apache Jena to tests
- add batch support for sparql insert
- Handle double quoted string content
- Handle newlines in string content
- http://iswc2011.semanticweb.org/fileadmin/iswc/Papers/Workshops/SSWS/Emmons-et-all-SSWS2011.pdf
had to be fixed
sqlite
sqlite outperforms all other approaches, especially if it used as an "in-memory" database. But even reading from disk is not much slower.
sqlite and JSON are similar in read and write performance. With sqlite SQL queries can be used wich is not directly possible with JSON.
Using SQL instead of JSON has some disadvantages e.g. a "schemaless" working mode is not possible. We worked around the problem by adding "ListOfDict" support where the columns are automatically detected and the DDL commands to create a schema need not be maintained manually.
Using SQL instead of a Graph database or Triplestore based approach has some disadvantages - e.g. graph queries are not directly possible but need to be translated to SQL - JOIN based relational queries. Given the low number of entities to be stored we hope this will not be too troublesome.
performance result for the "cities" example
see https://github.com/WolfgangFahl/DgraphAndWeaviateTest/blob/master/tests/testSqlite3.py
adding 128769 City records took 0.300 s => 428910 records/s selecting 128769 City records took 0.300 s => 428910 records/s