skip to main content

Search for: All records

Creators/Authors contains: "Niu, X."

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. We present an overview of GProM, a generic provenance middleware for relational databases. The sys- tem supports diverse provenance and annotation management tasks through query instrumentation, i.e., compiling a declarative frontend language with provenance-specific features into the query language of a backend database system. In addition to introducing GProM, we also discuss research contributions related to GProM including the first provenance model and capture mechanism for transaction prove- nance, a unified framework for answering why- and why-not provenance questions, and provenance- aware query optimization. Furthermore, by means of the example of post-mortem debugging of transac- tions, we demonstrate how novelmore »applications of provenance are made possible by GProM.« less
  2. The rigid schemas of classical relational databases help users in specifying queries and inform the storage organization of data. However, the advantages of schemas come at a high upfront cost through schema and ETL process design. In this work, we propose a new paradigm where the database system takes a more active role in schema development and data integration. We refer to this approach as adaptive schema databases (ASDs). An ASD ingests semi-structured or unstructured data directly using a pluggable combination of extraction and data integration techniques. Over time it discovers and adapts schemas for the ingested data using informationmore »provided by data integration and information extraction techniques, as well as from queries and user-feedback. In contrast to relational databases, ASDs maintain multiple schema workspaces that represent individualized views over the data, which are fine-tuned to the needs of a particular user or group of users. A novel aspect of ASDs is that probabilistic database techniques are used to encode ambiguity in automatically generated data extraction workflows and in generated schemas. ASDs can provide users with context-dependent feedback on the quality of a schema, both in terms of its ability to satisfy a user's queries, and the quality of the resulting answers. We outline our vision for ASDs, and present a proof of concept implementation as part of the Mimir probabilistic data curation system.« less
  3. Free, publicly-accessible full text available March 1, 2023
  4. Free, publicly-accessible full text available January 1, 2023
  5. A bstract A search for a heavy resonance decaying to a top quark and a W boson in the fully hadronic final state is presented. The analysis is performed using data from proton-proton collisions at a center-of-mass energy of 13 TeV, corresponding to an integrated luminosity of 137 fb − 1 recorded by the CMS experiment at the LHC. The search is focused on heavy resonances, where the decay products of each top quark or W boson are expected to be reconstructed as a single, large-radius jet with a distinct substructure. The production of an excited bottom quark, b *more », is used as a benchmark when setting limits on the cross section for a heavy resonance decaying to a top quark and a W boson. The hypotheses of b * quarks with left-handed, right-handed, and vector-like chiralities are excluded at 95% confidence level for masses below 2.6, 2.8, and 3.1 TeV, respectively. These are the most stringent limits on the b * quark mass to date, extending the previous best limits by almost a factor of two.« less
    Free, publicly-accessible full text available December 1, 2022
  6. Free, publicly-accessible full text available September 1, 2022
  7. Free, publicly-accessible full text available August 1, 2022
  8. Free, publicly-accessible full text available August 1, 2022
  9. A bstract Jet production in lead-lead (PbPb) and proton-proton (pp) collisions at a nucleon-nucleon center-of-mass energy of 5.02 TeV is studied with the CMS detector at the LHC, using PbPb and pp data samples corresponding to integrated luminosities of 404 μ b − 1 and 27.4 pb − 1 , respectively. Jets with different areas are reconstructed using the anti- k T algorithm by varying the distance parameter R . The measurements are performed using jets with transverse momenta ( p T ) greater than 200 GeV and in a pseudorapidity range of |η| < 2. To reveal the mediummore »modification of the jet spectra in PbPb collisions, the properly normalized ratio of spectra from PbPb and pp data is used to extract jet nuclear modification factors as functions of the PbPb collision centrality, p T and, for the first time, as a function of R up to 1.0. For the most central collisions, a strong suppression is observed for high- p T jets reconstructed with all distance parameters, implying that a significant amount of jet energy is scattered to large angles. The dependence of jet suppression on R is expected to be sensitive to both the jet energy loss mechanism and the medium response, and so the data are compared to several modern event generators and analytic calculations. The models considered do not fully reproduce the data.« less