Browse Definitions :

Analytics

This glossary contains definitions related to customer data analytics, predictive analytics, data visualization and operational business intelligence. Some definitions explain the meaning of words used to Hadoop and other software tools used in big data analytics. Other definitions are related to the strategies that business intelligence professionals, data scientists, statisticians and data analysts use to make data-driven decisions.

DAT - INF

  • data anonymization - The purpose of data anonymization is to make its source untraceable.
  • data breach - A data breach is a cyber attack in which sensitive, confidential or otherwise protected data has been accessed and/or disclosed in an unauthorized fashion.
  • data classification - Data classification is the process of organizing data into categories that make it is easy to retrieve, sort and store for future use.
  • data confabulation - Data confabulation is a business intelligence term for the selective and possibly misleading use of data to support a decision that has already been made.
  • data context - Data context is the network of connections among data points.
  • data democratization - Data democratization is the ability for information in a digital format to be accessible to the average end user.
  • data exhaust - Data exhaust is a byproduct of user actions online and consists of the various files generated by web browsers and their plug-ins such as cookies, log files, temporary internet files and and .
  • data federation software - Data federation software is programming that provides an organization with the ability to collect data from disparate sources and aggregate it in a virtual database where it can be used for business intelligence (BI) or other analysis.
  • data gravity - Data gravity is an attribute of data that is manifest in the way software and services are drawn to it relative to its mass (the amount of data).
  • data hygiene - Data hygiene is the collective processes conducted to ensure the cleanliness of data.
  • data integration - Data integration is the process of combining data from multiple source systems to create unified sets of information for both operational and analytical uses.
  • data journalism - Data journalism in an approach to writing for the public in which the journalist analyzes large data sets to identify potential news stories.
  • data latency - Data latency is the time it takes for data packets to be stored or retrieved.
  • data lifecycle management (DLM) - Data lifecycle management (DLM) is a policy-based approach to managing the flow of an information system's data throughout its lifecycle: from creation and initial storage to when it becomes obsolete and is deleted.
  • data literacy - Data literacy is the ability to derive information from data, just as literacy in general is the ability to derive information from the written word.
  • data marketplace (data market) - Data marketplaces typically offer various types of data for different markets and from different sources.
  • data masking - Data masking is a method of creating a structurally similar but inauthentic version of an organization's data that can be used for purposes such as software testing and user training.
  • data preparation - Data preparation is the process of gathering, combining, structuring and organizing data so it can be used in business intelligence (BI), analytics and data visualization applications.
  • data protection management (DPM) - Data protection management (DPM) comprises the administration, monitoring and management of backup processes to ensure backup tasks run on schedule and data is securely backed up and recoverable.
  • data sampling - Data sampling is a statistical analysis technique used to select, manipulate and analyze a representative subset of data points to identify patterns and trends in the larger data set being examined.
  • data science as a service (DSaaS) - Data science as a service (DSaaS) is a form of outsourcing that involves the delivery of information gleaned from advanced analytics applications run by data scientists at an outside company to corporate clients for their business use.
  • data science platform - A data science platform is software that allows data scientists to uncover actionable insights from data and communicate those insights throughout an enterprise within a single environment.
  • data scientist - A data scientist is a professional responsible for collecting, analyzing and interpreting extremely large amounts of data.
  • data smog - Data smog refers to the volume and velocity of data that is being created by devices connected to the Internet of Things.
  • data storytelling - Data storytelling is the process of translating complex data analyses into layman's terms in order to influence a decision or action.
  • data visualization - Data visualization is the practice of translating information into a visual context, such as a map or graph, to make data easier for the human brain to understand and pull insights from.
  • data-driven disaster - A data-driven disaster is a serious problem caused by one or more ineffective data analysis processes.
  • database management system (DBMS) - A database management system (DBMS) is system software for creating and managing databases.
  • database-agnostic - Database-agnostic is a term describing the capacity of software to function with any vendor’s database management system (DBMS).
  • de-anonymization (deanonymization) - De-anonymization is a method used to detect the original data that was subjected to processes to make it impossible -- or at least harder -- to identify the personally identifiable information (PII).
  • Decision Model and Notation (DMN) - Decision Model and Notation (DMN) is a formalized method of making and mapping out decisions through official business processes.
  • decision tree - A decision tree is a graph that uses a branching method to illustrate every possible output for a specific input.
  • decision-making process - The decision-making process, in a business context, is a set of steps taken by managers in an enterprise to determine the planned path for business initiatives and to set specific actions in motion.
  • deductive argument - A deductive argument is the presentation of statements that are assumed or known to be true as premises for a conclusion that necessarily follows from those statements.
  • deductive reasoning - Deductive reasoning is a logical process in which a conclusion is based on the accordance of multiple premises that are generally assumed to be true.
  • deep analytics - Deep analytics is the application of sophisticated data processing techniques to yield information from large and typically multi-source data sets comprised of both unstructured and semi-structured data.
  • deep learning - Deep learning is a type of machine learning and artificial intelligence (AI) that imitates the way humans gain certain types of knowledge.
  • demand planning - Demand planning is the process of forecasting the demand for a product or service so it can be produced and delivered more efficiently and to the satisfaction of customers.
  • Demand signal repository (DSR) - Demand signal repository (DSR) is a database that aggregates sales data at the point of sale (POS).
  • descriptive analytics - Descriptive analytics is a preliminary stage of data processing that creates a summary of historical data to yield useful information and possibly prepare the data for further analysis.
  • descriptive modeling - Descriptive modeling is a mathematical process that describes real-world events and the relationships between factors responsible for them.
  • direct digital marketing (DDM) - Direct digital marketing (DDM) is the electronic delivery of relevant communications to specific recipients.
  • document capture - Document capture is any one of several processes used to convert a physical document to another format, typically a digital representation.
  • driver-based planning - Driver-based planning is an approach to management that identifies an organization's key business drivers and creates a series of business plans that mathematically model how those things most necessary for the organization's success would be affected by different variables.
  • dynamic BPM (business process management) - Dynamic business process management (BPM) is an approach designed to allow business processes to adjust quickly to changing business needs.
  • dynamic case management (DCM) - Dynamic case management (DCM) is the handling of case-based work through the use of technologies that automate and streamline aspects of each case.
  • e-business (electronic business) - E-business (electronic business) is the conduct of business processes on the internet.
  • e-score - The e-score is a consumer rating metric used to to determine an individual's potential value as a customer and to use that information to guide marketing efforts.
  • econometrics - Econometrics is the analysis and testing of economic theories to verify hypotheses and improve prediction of financial trends.
  • edge analytics - Edge analytics is an approach to data collection and analysis in which an automated analytical computation is performed on data at a sensor, network switch or other device instead of waiting for the data to be sent back to a centralized data store.
  • Elastic - Elastic is a software company that provides products and services related to Elasticsearch, its distributed enterprise search engine.
  • Elastic Stack - Elastic Stack is a group of open source products from Elastic designed to help users take data from any type of source and in any format and search, analyze, and visualize that data in real time.
  • embedded BI (embedded business intelligence) - Embedded BI (business intelligence) is the integration of self-service BI tools into commonly used business applications.
  • emotional intelligence (EI) - Emotional intelligence (EI) is the area of cognitive ability that facilitates interpersonal behavior.
  • empiricism - Empiricism is the theory that human knowledge comes predominantly from experiences gathered through the five senses.
  • encryption key management - Encryption key management is the administration of tasks involved with protecting, storing, backing up and organizing encryption keys.
  • ensemble modeling - Ensemble modeling is the process of running two or more related but different analytical models and then synthesizing the results into a single score or spread in order to improve the accuracy of predictive analytics and data mining applications.
  • enterprise architecture (EA) - An enterprise architecture (EA) is a conceptual blueprint that defines the structure and operation of an organization.
  • enterprise document management (EDM) - Enterprise document management (EDM) is a strategy for overseeing an organization's paper and electronic documents so they can be easily retrieved in the event of a compliance audit or subpoena.
  • enterprise mashup (or data mashup) - An enterprise mashup is the integration of heterogeneous digital data and applications from multiple sources for business purposes.
  • enterprise search - There are a number of kinds of enterprise search including local installations, hosted versions, and search appliances, sometimes called “search in a box.
  • explicit data - Explicit data is data that is provided intentionally and taken at face value rather than analyzed or interpreted for further meaning.
  • Extract, Load, Transform (ELT) - Extract, Load, Transform (ELT) is a data integration process for transferring raw data from a source server to a data system (such as a data warehouse or data lake) on a target server and then preparing the information for downstream uses.
  • Fair Information Practices (FIP) - FIP (Fair Information Practices) is a general term for a set of standards governing the collection and use of personal data and addressing issues of privacy and accuracy.
  • falsifiability - Falsifiability is the capacity for some proposition, statement, theory or hypothesis to be proven wrong.
  • fast data - Fast data is the application of big data analytics to smaller data sets in near-real or real-time in order to solve a problem or create business value.
  • financial analytics - Financial analytics is the creation of ad hoc analysis to answer specific business questions and forecast possible future financial scenarios.
  • financial data management - Financial data management (FDM) is a process and policy, usually assisted by specialized software, that allows an enterprise or institution to consolidate its financial information, maintain compliance with accounting rules and laws, and produce detailed financial reports.
  • firmographic data - Firmographic data is types of information that can be used to categorize organizations, such as location, name, number of clients, industry and so on.
  • focus group - A focus group is a small and typically diverse panel of people selected to survey for their opinions on a given subject.
  • fresh data - Fresh data is data that is current and immediately usable and useful.
  • gap analysis - A gap analysis is a method of assessing the performance of a business unit to determine whether business requirements or objectives are being met and, if not, what steps should be taken to meet them.
  • Gartner - Gartner is an information technology (IT) research and consultancy company, formerly known as Gartner Group.
  • Gartner hype cycle - The hype cycle is a graphical representation of the life cycle stages a technology goes through from conception to maturity and widespread adoption.
  • Gartner Magic Quadrants - Gartner Magic Quadrants is a research method designed to monitor and evaluate the progress and positions of companies in a specific, technology-based market.
  • geolocation - Geolocation is the detection of the physical location of an Internet connected computing device.
  • Gmail Smart Compose - Gmail Smart Compose is a predictive text feature to help save effort in composing email messages.
  • GoodData - GoodData is a software company specializing in cloud-based business intelligence (BI) and big data analytics.
  • Google Analytics - Google Analytics is a free web analytics service that provides statistics and analytical tools for search engine optimization (SEO) and marketing purposes.
  • governance, risk management and compliance (GRC) - Governance, risk and compliance (GRC) refers to an organization's strategy for handling the interdependencies between corporate governance policies, enterprise risk management (ERM) programs, and regulatory and company compliance.
  • graph analytics - Graph analytics is a category of software tools and data mining techniques that help an analyst understand the relationship between entries in a graph database.
  • Hadoop - Hadoop is an open source distributed processing framework that manages data processing and storage for big data applications running in clustered systems.
  • Hadoop 2 - Apache Hadoop 2 is the second iteration of the Hadoop framework for distributed data processing.
  • Hadoop as a service (HaaS) - Hadoop as a service (HaaS), also known as Hadoop in the cloud, is a big data analytics framework that stores and analyzes data in the cloud using Hadoop.
  • halo effect - The halo effect is a form of cognitive bias which causes one part to make the whole seem more attractive or desirable.
  • health informatics - Health informatics is the practice of acquiring, studying and managing health data and applying medical concepts in conjunction with health information technology systems to help clinicians provide better healthcare.
  • heat map (heatmap) - A heat map is a two-dimensional representation of data in which values are represented by colors.
  • Hick's law - Hick's law is a psychological principle which states that the more options are available to a person, the longer it will take for him or her to make a decision about which option is best.
  • hypothesis - A hypothesis (plural: hypotheses) is a statement that may be tested and proven to be either true or false.
  • IBM Watson supercomputer - Watson is an IBM supercomputer that combines artificial intelligence (AI) and sophisticated analytical software for optimal performance as a 'question answering' machine.
  • ideation management - Ideation management refers to creating, gathering, analyzing, disseminating and executing on idea generation within the enterprise.
  • identity management (ID management) - Identity management (ID management) is the organizational process for ensuring that individuals have the appropriate access to technology resources.
  • implicit data - Implicit data is information that is not provided intentionally but gathered from available data streams.
  • imputation - Imputation, in statistics, is the insertion of a value to stand in for missing data.
  • in-database analytics - In-database analytics is a scheme for processing data within the database, avoiding the data movement that slows response time.
  • in-memory analytics - In-memory analytics queries data residing in a computer’s random access memory (RAM) rather than data stored on physical disks.
  • in-memory database - An in-memory database is a type of analytic database designed to streamline the work involved in processing queries.
  • incidence - Incidence, in statistics, is the rate of occurrence of something within a population, or the number of times it occurs.
  • inductive argument - An inductive argument is the use of collected instances of evidence of something specific to support a general conclusion.
  • inductive reasoning - Inductive reasoning is a logical process in which multiple premises, all believed true or found true most of the time, are combined to obtain a specific conclusion.
SearchCompliance
  • ISO 31000 Risk Management

    The ISO 31000 Risk Management framework is an international standard that provides businesses with guidelines and principles for ...

  • pure risk

    Pure risk refers to risks that are beyond human control and result in a loss or no loss with no possibility of financial gain.

  • risk reporting

    Risk reporting is a method of identifying risks tied to or potentially impacting an organization's business processes.

SearchSecurity
  • Twofish

    Twofish is a symmetric-key block cipher with a block size of 128 bits and variable-length key of size 128, 192 or 256 bits.

  • walled garden

    On the internet, a walled garden is an environment that controls the user's access to network-based content and services.

  • potentially unwanted program (PUP)

    A potentially unwanted program (PUP) is a program that may be unwanted, despite the possibility that users consented to download ...

SearchHealthIT
SearchDisasterRecovery
  • What is risk mitigation?

    Risk mitigation is a strategy to prepare for and lessen the effects of threats faced by a business.

  • fault-tolerant

    Fault-tolerant technology is a capability of a computer system, electronic system or network to deliver uninterrupted service, ...

  • synchronous replication

    Synchronous replication is the process of copying data over a storage area network, local area network or wide area network so ...

SearchStorage
  • Remote Direct Memory Access (RDMA)

    Remote Direct Memory Access (RDMA) is a technology that enables two networked computers to exchange data in main memory without ...

  • storage (computer storage)

    Data storage is the collective methods and technologies that capture and retain digital information on electromagnetic, optical ...

  • storage medium (storage media)

    In computers, a storage medium is a physical device that receives and retains electronic data for applications and users and ...

Close