Data Repository

14,000,000 Leading Edge Experts on the ideXlab platform

Scan Science and Technology

Contact Leading Edge Experts & Companies

Scan Science and Technology

Contact Leading Edge Experts & Companies

The Experts below are selected from a list of 100356 Experts worldwide ranked by ideXlab platform

Akira Kato - One of the best experts on this subject based on the ideXlab platform.

  • USENIX Annual Technical Conference, FREENIX Track - Traffic Data Repository at the WIDE project
    2000
    Co-Authors: Kenjiro Cho, Koushirou Mitsuya, Akira Kato
    Abstract:

    It becomes increasingly important for both network researchers and operators to know the trend of network traffic and to find anomaly in their network traffic. This paper describes an on-going effort within the WIDE project to collect a set of free tools to build a traffic Data Repository containing detailed information of our backbone traffic. Traffic traces are collected by tcpdump and, after removing privacy information, the traces are made open to the public. We review the issues on user privacy, and then, the tools used to build the WIDE traffic Repository. We will report the current status and findings in the early stage of our IPv6 deployment.

  • traffic Data Repository at the wide project
    USENIX Annual Technical Conference, 2000
    Co-Authors: Kenjiro Cho, Koushirou Mitsuya, Akira Kato
    Abstract:

    It becomes increasingly important for both network researchers and operators to know the trend of network traffic and to find anomaly in their network traffic. This paper describes an on-going effort within the WIDE project to collect a set of free tools to build a traffic Data Repository containing detailed information of our backbone traffic. Traffic traces are collected by tcpdump and, after removing privacy information, the traces are made open to the public. We review the issues on user privacy, and then, the tools used to build the WIDE traffic Repository. We will report the current status and findings in the early stage of our IPv6 deployment.

Ladjel Bellatreche - One of the best experts on this subject based on the ideXlab platform.

  • The role of user requirements in Data Repository design
    International Journal on Software Tools for Technology Transfer, 2018
    Co-Authors: Ilyes Boukhari, Idir Ait-sadoune, Stéphane Jean, Ladjel Bellatreche
    Abstract:

    Requirements engineering plays a crucial role in the development process of an information system as it aims at providing a complete and accurate requirement specification. In the life cycle of a Data Repository ($${\mathcal {D}}{\mathcal {R}}$$DR) such as a Database or a Data warehouse, the requirements are mainly used to define the conceptual model once they have been identified from the informal specification. In this paper, we study the interest of requirements in the other phases of the $${\mathcal {D}}{\mathcal {R}}$$DR life cycle. As the Data integration problem, handled in the Extract, Transform, Load (ETL) phase, comes from the heterogeneity of requirements, we introduce a requirement integration framework based on ontologies and a generic model to unify the used vocabularies and requirement languages. Then we propose an approach to check the consistency of the requirements, w.r.t. the integrity constraints defined on the logical schema using the formal B method. We also show that requirements help define appropriate access structures such as indexes and materialized views to optimize SQL queries of a $${\mathcal {D}}{\mathcal {R}}$$DR. Our approach is based on transformation rules that identify important queries that will be executed on a $${\mathcal {D}}{\mathcal {R}}$$DR directly from the requirements. The experiments conducted on the Star Schema Benchmark (SSB) confirm the interest of this approach for the selection of different optimization structures. Finally, we present the OntoReqTool that implements the previous functionality on top of the OntoDB/OntoQL platform.

  • On Using Requirements Throughout the Life Cycle of Data Repository
    2014
    Co-Authors: Stéphane Jean, Ladjel Bellatreche, Idir Ait-sadoune, Ilyes Boukhari
    Abstract:

    Requirements engineering aims at providing a requirement specification with some nice properties such as completeness or accuracy. In the lifecycle of a Data Repository ( DR ), user requirements are usually assumed to be homogenous and used mainly to define the conceptual model of a DR . In this paper, we study the interest of the requirements in the other phases of the life cycle of a DR . We propose a generic model based on ontologies to unify the used vocabularies and requirements languages. Then we extend this model using the formal method B to check the consistency of the requirements w.r.t. the integrity constraints defined on the logical schema. Finally we propose to select optimization structures of a DR using the user requirements instead of SQL queries. Several experiments on the Star Schema Benchmark (SSB) confirm the interest of our proposition.

Kenjiro Cho - One of the best experts on this subject based on the ideXlab platform.

  • USENIX Annual Technical Conference, FREENIX Track - Traffic Data Repository at the WIDE project
    2000
    Co-Authors: Kenjiro Cho, Koushirou Mitsuya, Akira Kato
    Abstract:

    It becomes increasingly important for both network researchers and operators to know the trend of network traffic and to find anomaly in their network traffic. This paper describes an on-going effort within the WIDE project to collect a set of free tools to build a traffic Data Repository containing detailed information of our backbone traffic. Traffic traces are collected by tcpdump and, after removing privacy information, the traces are made open to the public. We review the issues on user privacy, and then, the tools used to build the WIDE traffic Repository. We will report the current status and findings in the early stage of our IPv6 deployment.

  • traffic Data Repository at the wide project
    USENIX Annual Technical Conference, 2000
    Co-Authors: Kenjiro Cho, Koushirou Mitsuya, Akira Kato
    Abstract:

    It becomes increasingly important for both network researchers and operators to know the trend of network traffic and to find anomaly in their network traffic. This paper describes an on-going effort within the WIDE project to collect a set of free tools to build a traffic Data Repository containing detailed information of our backbone traffic. Traffic traces are collected by tcpdump and, after removing privacy information, the traces are made open to the public. We review the issues on user privacy, and then, the tools used to build the WIDE traffic Repository. We will report the current status and findings in the early stage of our IPv6 deployment.

A Galyam - One of the best experts on this subject based on the ideXlab platform.

  • wiserep an interactive supernova Data Repository
    Publications of the Astronomical Society of the Pacific, 2012
    Co-Authors: Ofer Yaron, A Galyam
    Abstract:

    We have entered an era of massive Data sets in astronomy. In particular, the number of supernova (SN) discoveries and classifications has substantially increased over the years from few tens to thousands per year. It is no longer the case that observations of a few prototypical events encapsulate most spectroscopic information about SNe, motivating the development of modern tools to collect, archive, organize, and distribute spectra in general and SN spectra in particular. For this reason, we have developed the Weizmann Interactive Supernova Data Repository (WISeREP)—an SQL-based Database (DB) with an interactive Web-based graphical interface. The system serves as an archive of high-quality SN spectra, including both historical (legacy) Data and Data that are accumulated by ongoing modern programs. The archive provides information about objects, their spectra, and related metaData. Utilizing interactive plots, we provide a graphical interface to visualize Data, perform line identification of the major relevant species, determine object redshifts, classify SNe, and measure expansion velocities. Guest users may view and download spectra or other Data that have been placed in the public domain. Registered users may also view and download Data that are proprietary to specific programs with which they are associated. The DB currently holds more than 8000 spectra, of which more than 5000 are public; the latter include published spectra from the Palomar Transient Factory (PTF), all of the SUSPECT (Supernova Spectrum) archive, the Caltech-Core-Collapse Program (CCCP), the CfA SN spectra archive, and published spectra from the University of California, Berkeley, SNDB Repository. It offers an efficient and convenient way to archive Data and share it with colleagues, and we expect that Data stored in this way will be easy to access, increasing its visibility, usefulness, and scientific impact. We encourage the SN community worldwide to make use of the Data and tools provided by WISeREP and to contribute Data to be made globally available and archived for posterity.

  • wiserep an interactive supernova Data Repository
    arXiv: Instrumentation and Methods for Astrophysics, 2012
    Co-Authors: Ofer Yaron, A Galyam
    Abstract:

    We have entered an era of massive Data sets in astronomy. In particular, the number of supernova (SN) discoveries and classifications has substantially increased over the years from few tens to thousands per year. It is no longer the case that observations of a few prototypical events encapsulate most spectroscopic information about SNe, motivating the development of modern tools to collect, archive, organize and distribute spectra in general, and SN spectra in particular. For this reason we have developed the Weizmann Interactive Supernova Data Repository - WISeREP - an SQL-based Database (DB) with an interactive web-based graphical interface. The system serves as an archive of high quality SN spectra, including both historical (legacy) Data as well as Data that is accumulated by ongoing modern programs. The archive provides information about objects, their spectra, and related meta-Data. Utilizing interactive plots, we provide a graphical interface to visualize Data, perform line identification of the major relevant species, determine object redshifts, classify SNe and measure expansion velocities. Guest users may view and download spectra or other Data that have been placed in the public domain. Registered users may also view and download Data that are proprietary to specific programs with which they are associated. The DB currently holds >8000 spectra, of which >5000 are public; the latter include published spectra from the Palomar Transient Factory, all of the SUSPECT archive, the Caltech-Core-Collapse Program, the CfA SN spectra archive and published spectra from the UC Berkeley SNDB Repository. It offers an efficient and convenient way to archive Data and share it with colleagues, and we expect that Data stored in this way will be easy to access, increasing its visibility, usefulness and scientific impact.

Ilyes Boukhari - One of the best experts on this subject based on the ideXlab platform.

  • The role of user requirements in Data Repository design
    International Journal on Software Tools for Technology Transfer, 2018
    Co-Authors: Ilyes Boukhari, Idir Ait-sadoune, Stéphane Jean, Ladjel Bellatreche
    Abstract:

    Requirements engineering plays a crucial role in the development process of an information system as it aims at providing a complete and accurate requirement specification. In the life cycle of a Data Repository ($${\mathcal {D}}{\mathcal {R}}$$DR) such as a Database or a Data warehouse, the requirements are mainly used to define the conceptual model once they have been identified from the informal specification. In this paper, we study the interest of requirements in the other phases of the $${\mathcal {D}}{\mathcal {R}}$$DR life cycle. As the Data integration problem, handled in the Extract, Transform, Load (ETL) phase, comes from the heterogeneity of requirements, we introduce a requirement integration framework based on ontologies and a generic model to unify the used vocabularies and requirement languages. Then we propose an approach to check the consistency of the requirements, w.r.t. the integrity constraints defined on the logical schema using the formal B method. We also show that requirements help define appropriate access structures such as indexes and materialized views to optimize SQL queries of a $${\mathcal {D}}{\mathcal {R}}$$DR. Our approach is based on transformation rules that identify important queries that will be executed on a $${\mathcal {D}}{\mathcal {R}}$$DR directly from the requirements. The experiments conducted on the Star Schema Benchmark (SSB) confirm the interest of this approach for the selection of different optimization structures. Finally, we present the OntoReqTool that implements the previous functionality on top of the OntoDB/OntoQL platform.

  • On Using Requirements Throughout the Life Cycle of Data Repository
    2014
    Co-Authors: Stéphane Jean, Ladjel Bellatreche, Idir Ait-sadoune, Ilyes Boukhari
    Abstract:

    Requirements engineering aims at providing a requirement specification with some nice properties such as completeness or accuracy. In the lifecycle of a Data Repository ( DR ), user requirements are usually assumed to be homogenous and used mainly to define the conceptual model of a DR . In this paper, we study the interest of the requirements in the other phases of the life cycle of a DR . We propose a generic model based on ontologies to unify the used vocabularies and requirements languages. Then we extend this model using the formal method B to check the consistency of the requirements w.r.t. the integrity constraints defined on the logical schema. Finally we propose to select optimization structures of a DR using the user requirements instead of SQL queries. Several experiments on the Star Schema Benchmark (SSB) confirm the interest of our proposition.