Browse Definitions :

Data and data management

Terms related to data, including definitions about data warehousing and words and phrases about data management.

3-T - DAT

  • 3-tier application architecture - A 3-tier application architecture is a modular client-server architecture that consists of a presentation tier, an application tier and a data tier.
  • 3Vs (volume, variety and velocity) - The 3Vs (volume, variety and velocity) are three defining properties or dimensions of big data.
  • 5 V's of big data - The 5 V's of big data are the five main and innate characteristics of big data.
  • accounting error - What is an accounting error?An accounting error is a non-fraudulent discrepancy in financial documentation.
  • active archive - An active archive is a collection of data that is too valuable for a company to discard, but only needs to be accessed occasionally.
  • ActiveX Data Objects (ADO) - ActiveX Data Objects (ADO) is an application program interface from Microsoft that lets a programmer writing Windows applications get access to a relational or non-relational database from both Microsoft and other database providers.
  • Adaptive Server Enterprise (ASE) - Adaptive Server Enterprise (ASE) is a relational database management system (RDBMS) from Sybase, Inc.
  • Allscripts - Allscripts is a vendor of electronic health record systems for physician practices, hospitals and healthcare systems.
  • alternate data stream (ADS) - An alternate data stream (ADS) is a feature of Windows New Technology File System (NTFS) that contains metadata for locating a specific file by author or title.
  • alternative data - Alternative data is information gathered from non-traditional information sources.
  • Amalga - Amalga is a health IT data integration platform designed to retrieve and display healthcare-related information from various sources, including scanned documents, lab results, dictated notes and images such as X-rays, EKGs and MRIs.
  • Amazon Kinesis - Amazon Kinesis is the fully managed Amazon Web Service (AWS) offering for real-time processing of big data.
  • Amazon Redshift - Amazon Redshift is a fully managed petabyte-scale data warehouse service.
  • Amazon Simple Storage Service (Amazon S3) - Amazon Simple Storage Service (Amazon S3) is a scalable, high-speed, web-based cloud storage service.
  • Anaplan - Anaplan is a Web-based enterprise platform for business planning.
  • Apache Falcon - Apache Falcon is a data management tool for overseeing data pipelines in Hadoop clusters, with a goal of ensuring consistent and dependable performance on complex processing jobs.
  • Apache Giraph - Apache Giraph is real-time graph processing software that is mostly used to analyze social media data.
  • Apache Hive - Apache Hive is an open source data warehouse system for querying and analyzing large data sets that are principally stored in Hadoop files.
  • Apache Parquet - Apache Parquet is a column-oriented storage format for Hadoop.
  • Apache Solr - Apache Solr is an open source search platform built upon a Java library called Lucene.
  • Apache Storm - Storm is a free and open source (FOSS) distributed real-time computation system developed by the Apache Foundation.
  • API endpoint - An API endpoint is a point at which an API -- the code that allows two software programs to communicate with each other -- connects with the software program.
  • Apple User Enrollment - Apple User Enrollment (UE) is a form of mobile device management (MDM) for Apple products that supports iOS 13 and macOS Catalina.
  • archive - An archive is a collection of data moved to a repository for long-term retention, to keep separate for compliance reasons or for moving off primary storage media.
  • AS1 (Applicability Statement 1) - AS1 (Applicability Statement is a specification for Electronic Data Interchange (EDI) communications between businesses using e-mail protocols.
  • atomic data - In a data warehouse, atomic data is the lowest level of detail.
  • AUFS (Advanced Multi-Layered Unification Filesystem) - Advanced multi-layered unification filesystem (AUFS) is a union filesystem sometimes used in platform-as-a-service environments to merge distinct directory hierarchies into a single directory.
  • autonomous transaction - In Oracle's database products, an autonomous transaction is an independent transaction that is initiated by another transaction.
  • Azure Data Studio (formerly SQL Operations Studio) - Azure Data Studio is a Microsoft tool, originally named SQL Operations Studio, for managing SQL Server databases and cloud-based Azure SQL Database and Azure SQL Data Warehouse systems.
  • B-tree - A B-tree is a method of placing and locating files (called records or keys) in a database.
  • bar graph - A bar graph is a pictorial rendition of statistical data in which the independent variable can attain only certain discrete values.
  • Bayesian statistics - Bayesian statistics is a mathematical approach to calculating probability in which conclusions are subjective and updated as additional data is collected.
  • big data - Big data is a combination of structured, semistructured and unstructured data collected by organizations that can be mined for information and used in machine learning projects, predictive modeling and other advanced analytics applications.
  • big data analytics - Big data analytics is the often complex process of examining big data to uncover information -- such as hidden patterns, correlations, market trends and customer preferences -- that can help organizations make informed business decisions.
  • big data as a service (BDaaS) - Big data as a service (BDaaS) is the delivery of data platforms and tools by a cloud provider to help organizations process, manage and analyze large data sets so they can generate insights in order to improve business operations and gain a competitive advantage.
  • big data engineer - A big data engineer is an information technology (IT) professional who is responsible for designing, building, testing and maintaining complex data processing systems that work with large data sets.
  • big data management - Big data management is the organization, administration and governance of large volumes of both structured and unstructured data.
  • big data storage - Big data storage is a compute-and-storage architecture that collects and manages large data sets and enables real-time data analytics.
  • binary tree - A binary tree is a method of placing and locating files (called records or keys) in a database, especially when all the data is known to be in random access memory (RAM).
  • bit rot - Bit rot is the slow deterioration in the performance and integrity of data stored on storage media.
  • BLOB (binary large object) - In computers, a BLOB (binary large object), pronounced BLAHB and sometimes spelled in all lower case, is a large file, typically an image or sound file, that must be handled (for example, uploaded, downloaded, or stored in a database) in a special way because of its size.
  • block - A block is a contiguous set of bits or bytes that forms an identifiable unit of data.
  • block diagram - A block diagram is a visual representation of a system that uses simple, labeled blocks that represent single or multiple items, entities or concepts, connected by lines to show relationships between them.
  • blockchain - Blockchain is a record-keeping technology designed to make it impossible to hack the system or forge the data stored on it, thereby making it secure and immutable.
  • blockchain storage - Blockchain storage is a way of saving data in a decentralized network, which utilizes the unused hard disk space of users across the world to store files.
  • Blue Cloud - Blue Cloud is an approach to shared infrastructure developed by IBM.
  • box plot - A box plot is a graphical rendition of statistical data based on the minimum, first quartile, median, third quartile, and maximum.
  • brontobyte - A brontobyte is a measure of memory or data storage that is equal to 10 to the 27th power of bytes.
  • business intelligence dashboard - A business intelligence dashboard, or BI dashboard, is a data visualization and analysis tool that displays on one screen the status of key performance indicators (KPIs) and other important business metrics and data points for an organization, department, team or process.
  • capacity management - Capacity management is the broad term describing a variety of IT monitoring, administration and planning actions that are taken to ensure that a computing infrastructure has adequate resources to handle current data processing requirements as well as the capacity to accommodate future loads.
  • case - A case is a particular instance of something.
  • catalog - In computing, a catalog is a directory of information about data sets, files, or a database.
  • causation - Causation, or causality, is the capacity of one variable to influence another.
  • CCHIT - Certification Commission for Healthcare Information Technology - The Certification Commission for Healthcare Information Technology (CCHIT) is an independent, not-for-profit group that certifies electronic health records (EHR) and networks for health information exchange (HIE) in the United States.
  • change data capture (CDC) - Change data capture (CDC) is the process of capturing changes made at the data source and applying them throughout the enterprise.
  • chief data officer (CDO) - A chief data officer (CDO) is a C-level executive who is responsible for an organization's data use and data governance.
  • CICS (Customer Information Control System) - CICS (Customer Information Control System) is middleware that sits between the z/OS IBM mainframe operating system and business applications.
  • clinical decision support system (CDSS) - A clinical decision support system (CDSS) is an application that analyzes data to help healthcare providers make decisions and improve patient care.
  • cloud audit - A cloud audit is a periodic examination an organization does to assess and document its cloud vendor's performance.
  • Cloud Data Management Interface - Cloud Data Management Interface (CDMI) is a Storage Networking Industry Association (SNIA) industry standard that defines the interface that applications will use to create, retrieve, update and delete data elements from the cloud.
  • cloud disaster recovery (cloud DR) - Cloud disaster recovery (cloud DR) is a combination of strategies and services intended to back up data, applications and other resources to public cloud or dedicated service providers.
  • cloud SLA (cloud service-level agreement) - A cloud SLA (cloud service-level agreement) is an agreement between a cloud service provider and a customer that ensures a minimum level of service is maintained.
  • cloud storage - Cloud storage is a service model in which data is transmitted and stored on remote storage systems, where it is maintained, managed, backed up and made available to users over a network (typically the internet).
  • cloud storage API - A cloud storage API is an application program interface that connects a locally-based application to a cloud-based storage system, so that a user can send data to it and access and work with data stored in it.
  • cloud storage service - A cloud storage service is a business that maintains and manages its customers' data and makes that data accessible over a network, usually the internet.
  • cohort - A cohort is a group of people that have some demographic or statistical characteristic in common.
  • cold backup (offline backup) - A cold backup, also called an offline backup, is a database backup during which the database is offline and not accessible to update.
  • Collaboration Data Objects (CDO) - Collaboration Data Objects (CDO) is Microsoft's technology for building messaging or collaboration applications or adding these capabilities to existing applications.
  • column database management system (CDBMS) - There are different types of CDBMS offerings, with the common defining feature being that data is stored by column (or column families) instead of as rows.
  • comma-separated values file (CSV) - In computers, a CSV (comma-separated values) file contains the values in a table as a series of ASCII text lines organized so that each column value is separated by a comma from the next column's value and each row starts a new line.
  • compliance as a service (CaaS) - Compliance as a Service (CaaS) is a cloud service service level agreement (SLA) that specified how a managed service provider (MSP) will help an organization meet its regulatory compliance mandates.
  • confidentiality - Confidentiality is a set of rules or a promise that limits access or places restrictions on certain types of information.
  • conformed dimension - In data warehousing, a conformed dimension is a dimension that has the same meaning to every fact with which it relates.
  • consensus algorithm - A consensus algorithm is a process in computer science used to achieve agreement on a single data value among distributed processes or systems.
  • consumer data - Consumer data is the information trail customers leave behind as a result of their Internet use.
  • content management application (CMA) - A content management application (CMA) is the front end component of a content management system (CMS).
  • content personalization - Content personalization is a strategy that tailors webpages and other forms of content to individual users' characteristics or preferences.
  • contiguous - Contiguous describes two or more objects that are adjacent to each other.
  • continuum - A continuum is a continuous system or range in which adjacent elements do not vary from each other in any marked degree although the endpoints of the system may be drastically different.
  • cooked data - Cooked data is raw data after it has been processed - that is, extracted, organized, and perhaps analyzed and presented - for further use.
  • copy data - Copy data is the electronic data that is created as a result of data protection functions like backups, snapshots and disaster recovery.
  • core banking system - A core banking system is the software used to support a bank’s most common transactions.
  • correlation - Correlation is a statistical measure that indicates the extent to which two or more variables fluctuate in relation to each other.
  • correlation coefficient - A correlation coefficient is a statistical measure of the degree to which changes to the value of one variable predict change to the value of another.
  • CouchDB - CouchDB is an open source document-oriented database based on common web standards.
  • CRM analytics - CRM (customer relationship management) analytics comprises all programming that analyzes data about customers and presents it to help facilitate and streamline better business decisions.
  • CRUD cycle (Create, Read, Update and Delete Cycle) - The CRUD cycle describes the elemental functions of a persistent database in a computer.
  • cryptographic nonce - A nonce is a random or semi-random number that is generated for a specific use.
  • curation - Curation is a field of endeavor involved with assembling, managing and presenting some type of collection.
  • customer data integration (CDI) - Customer data integration (CDI) is the process of defining, consolidating and managing customer information across an organization's business units and systems to achieve a "single version of the truth" for customer data.
  • customer intelligence (CI) - Customer intelligence (CI) is the process of gathering and analyzing information derived from customer data that an organization collects from both internal and external sources.
  • customer segmentation - Customer segmentation is the practice of dividing a customer base into groups of individuals that are similar in specific ways relevant to marketing, such as age, gender, interests and spending habits.
  • DAO (Data Access Objects) - DAO (Data Access Objects) is an application program interface (API) available with Microsoft's Visual Basic that lets a programmer request access to a Microsoft Access database.
  • dark data - Dark data is digital information an organization collects, processes and stores that is not currently being used for business purposes.
SearchCompliance
  • ISO 31000 Risk Management

    The ISO 31000 Risk Management framework is an international standard that provides businesses with guidelines and principles for ...

  • pure risk

    Pure risk refers to risks that are beyond human control and result in a loss or no loss with no possibility of financial gain.

  • risk reporting

    Risk reporting is a method of identifying risks tied to or potentially impacting an organization's business processes.

SearchSecurity
  • Twofish

    Twofish is a symmetric-key block cipher with a block size of 128 bits and variable-length key of size 128, 192 or 256 bits.

  • walled garden

    On the internet, a walled garden is an environment that controls the user's access to network-based content and services.

  • potentially unwanted program (PUP)

    A potentially unwanted program (PUP) is a program that may be unwanted, despite the possibility that users consented to download ...

SearchHealthIT
SearchDisasterRecovery
  • What is risk mitigation?

    Risk mitigation is a strategy to prepare for and lessen the effects of threats faced by a business.

  • fault-tolerant

    Fault-tolerant technology is a capability of a computer system, electronic system or network to deliver uninterrupted service, ...

  • synchronous replication

    Synchronous replication is the process of copying data over a storage area network, local area network or wide area network so ...

SearchStorage
  • hard disk drive (HDD)

    A computer hard disk drive (HDD) is a non-volatile data storage device.

  • Remote Direct Memory Access (RDMA)

    Remote Direct Memory Access (RDMA) is a technology that enables two networked computers to exchange data in main memory without ...

  • storage (computer storage)

    Data storage is the collective methods and technologies that capture and retain digital information on electromagnetic, optical ...

Close