Development of a semi-synthetic dataset as a testbed for big-data semantic analytics

Robert Techentin, Daniel Foti, Peter Li, Erik Daniel, Barry Gilbert, David Holmes, Sinan Al-Saffar

Research output: Chapter in Book/Report/Conference proceedingConference contribution

1 Scopus citations

Abstract

We have developed a large semi-synthetic, semantically rich dataset, modeled after the medical record of a large medical institution. Using the highly diverse data.gov data repository and a multivariate data augmentation strategy, we can generate arbitrarily large semi-synthetic datasets which can be used to test new algorithms and computational platforms. The construction process and basic data characterization are described. The databases, as well as code for data collection, consolidation, and augmentation are available for distribution.

Original languageEnglish (US)
Title of host publicationProceedings - 2014 IEEE International Conference on Semantic Computing, ICSC 2014
PublisherIEEE Computer Society
Pages252-253
Number of pages2
ISBN (Print)9781479940028
DOIs
StatePublished - 2014
Event8th IEEE International Conference on Semantic Computing, ICSC 2014 - Newport Beach, CA, United States
Duration: Jun 16 2014Jun 18 2014

Publication series

NameProceedings - 2014 IEEE International Conference on Semantic Computing, ICSC 2014

Other

Other8th IEEE International Conference on Semantic Computing, ICSC 2014
Country/TerritoryUnited States
CityNewport Beach, CA
Period6/16/146/18/14

Keywords

  • RDF
  • big data
  • data.gov
  • graph computing
  • semantic representation

ASJC Scopus subject areas

  • Software

Fingerprint

Dive into the research topics of 'Development of a semi-synthetic dataset as a testbed for big-data semantic analytics'. Together they form a unique fingerprint.

Cite this