Quality / governance Definitions

  • A

    Apache Falcon

    Apache Falcon is a data management tool for overseeing data pipelines in Hadoop clusters, with a goal of ensuring consistent and dependable performance on complex processing jobs.

  • C

    cooked data

    Cooked data is raw data after it has been processed - that is, extracted, organized, and perhaps analyzed and presented - for further use.

  • D

    data

    In computing, data is information that has been translated into a form that is efficient for movement or processing.

  • data analytics (DA)

    Data analytics (DA) is the science of examining raw data with the purpose of drawing conclusions about that information.

  • data governance (DG)

    Data governance (DG) is the overall management of the availability, usability, integrity and security of data used in an enterprise. A sound data governance program includes a governing body or council, a defined set of procedures and a plan to execute those procedures.

  • data management

    Data management is the practice of organizing and maintaining data processes to meet ongoing information lifecycle needs.

  • data modeling

    Data modeling is the process of documenting a complex software system design as an easily understood diagram, using text and symbols to represent the way data needs to flow.

  • data profiling

    Data profiling is the process of examining, analyzing and reviewing data to collect statistics surrounding the quality and hygiene of the dataset.

  • data quality

    Data quality is a perception or an assessment of data's fitness to serve its purpose in a given context.

  • data scrubbing (data cleansing)

    Data scrubbing, also called data cleansing, is the process of cleaning up data in a database that is incorrect, incomplete, or duplicated.

  • data stewardship

    Data stewardship is the management and oversight of an organization's data assets to provide business users with high quality data that is easily accessible in a consistent manner.

  • DataOps (data operations)

    DataOps (data operations) is an approach to designing, implementing and maintaining a distributed data architecture that will support a wide range of open source tools and frameworks in production.

  • dirty data

    In a data warehouse, dirty data is a database record that contains errors.

  • disambiguation

    Disambiguation (also called word sense disambiguation) is the act of interpreting the intended sense or meaning of a word. Disambiguation is a common problem in computer language processing, since it is often difficult for a computer to distinguish a word’s sense when the word has multiple meanings or spellings.

  • F

    fixed data (permanent data, reference data, archival data, or fixed-content data)

    Fixed data (sometimes referred to as permanent data) is data that is not, under normal circumstances, subject to change. Any type of historical record is fixed data. For example, meteorological details for a given location on a specific day in the past are not likely to change (unless the original record is found, somehow, to be flawed).

-ADS BY GOOGLE

SearchBusinessAnalytics

SearchAWS

SearchContentManagement

SearchOracle

SearchSAP

SearchSQLServer

Close