Data Access Management
Self Service Business Intelligence
External Data Sharing
Book a demo
Start for free
Modern Data Glossary
Find, understand and get familiar with data terms.
What is a data product manager?
Data Product Managers are people who understand business, technology and how to work with people to deliver a great product
What is rdbms
Relational database management systems or RDBMS are the most common type of database management system
What is a Recommendation Engine?
Recommendation engines are typically used to predict the rating or preference that a user would give to an item.
What is DataOps?
DataOps is the management of data integration. It's the combination of data, software engineering, and operations.
What is Data Stewardship?
Data stewardship is the set of people, processes and tools that ensure consistent and reliable access to data
What is Data Validation?
Data validation is a key aspect of software quality assurance and represents a software development process where data is tested to ensure it meets certain criteria
What are Data Silos?
Data silos create duplicate work, waste time and money, and prevent organizations from learning from their own experiences.
What is Data Fabric?
Data Fabric is software architecture for data management that unifies and integrates data across multiple systems
What is Data Confidentiality?
Data confidentiality is a set of rules or a promise that limits access or places restrictions on any information that is being shared. Data confidentiality is a component of information security and privacy.
What is Anonymized Data?
An anonymized dataset is a dataset in which personally identifiable information has been replaced by artificial identifiers, or masked.
What is an ETL pipeline?
Extract, transform and load (ETL) is a process in data warehousing responsible for pulling data out of the source systems and placing it into a data warehouse.
What is Master Data Management?
MDM is a way to bring together critical business data in one place and help ensure that it's accurate, consistent, complete and trusted across the enterprise
What is Data Preparation?
Data preparation is sometimes referred to as "data wrangling." It's the process of taking raw data from a variety of sources and getting it ready for analysis
What is Deduplication?
Data deduplication is a process that identifies and removes duplicate copies of repeating data segments to free up storage capacity and reduce costs
What is Data Wrangling?
Data wrangling is the process of transforming and mapping data from one "raw" data form into another format with the intent of making it more appropriate and valuable for a variety of downstream purposes
What is Data Ingestion?
Data ingestion is the transportation of data from assorted sources to a storage medium where it can be accessed and analyzed by an enterprise
What is Data Architecture?
Data architecture is the design of data for use in defining the target state and the subsequent planning needed to achieve the target state
What is Apache Airflow?
Apache Airflow is a platform to programmatically author, schedule and monitor workflows
What is Data Analytics?
Data analytics is an umbrella term for a number of different ways that data can be analyzed.
What is Data Privacy?
Data privacy is the right of people to control their own personal data.
What is a Relational Database?
A relational database is a type of database that stores and provides access to data points that are related to one another.
What is a Metrics Layer?
Metrics layer is a type of software that allows you to access data from multiple sources, transform the data and send it to another system
What is Data Modelling?
Data Modelling is a method used to define and analyze data requirements needed to support the business processes within the scope of corresponding information systems in organizations
What is a Data Model?
A data model is a set of concepts for describing data, data relationships, data semantics and constraints
What is Data Discovery?
Data discovery is also known as “data exploration” as it involves exploratory analysis of big datasets with no prior knowledge of what results you may expect from the analysis
What is a Data Platform?
A data platform is a central repository of enterprise data from which data scientists and business analysts can derive value
What is a Data Analyst?
Data analysts are the people who take that data and use it to help companies make better business decisions
What is Business Intelligence?
BI technologies provide historical, current and predictive views of business operations
What is a Data Governance Framework?
A data governance framework is an organizational strategy that defines who is responsible for managing and protecting information collected by a company or organization
What is a Data Catalog?
A data catalog is a metadata management tool designed to help users discover and understand information about data
What is a Query?
The most common type of SQL query is a SELECT query, which retrieves data from one or more tables in your database and returns it in the form of one or more rows.
What is a Primary Key?
A primary key is simply a unique identifier for a row.
What is Metabase?
Metabase is the easy, open source way for everyone in your company to ask questions and learn from data
What is Data Integrity?
Data obfuscating is the maintenance, assurance and protection of sensitive information in your database
What is a Data Engineer?
Data engineers provide this access by building the architecture necessary to store, process and analyze data
What is Redash?
Redash features a simple interface that allows you to create new queries, visualize the data and share dashboards
What is Data Curation?
Curation allows users to find the right data at the right time and ensures that it remains usable in the future
What is Metadata Management?
Metadata also lets you create a more secure database by making it easier for you to identify where sensitive information is located within your system
What is SQL?
SQL (pronounced "ess-que-el") stands for Structured Query Language
What is a Data Lake?
A data lake is a repository that holds a vast amount of raw data in its native format until it is needed.
What is Data Profiling?
Data profiling is a process that discovers, analyzes and displays the content, structure, quality and consistency of a given set of data
What is an Entity Relationship Diagram?
An Entity Relationship Diagram (ERD) is a visual representation of different data using conventions that describe how these data are related to each other
What is Data Lineage?
Data lineage enables you to trace the path of a specific piece of data as it moves throughout your data ecosystem
What is Metadata?
Metadata is information that describes and gives information about other pieces of data.
What is Data Governance?
Data governance is a set of activities that oversee the quality of information used for critical decision making.
What is Monthly Recurring Revenue?
Monthly recurring revenue (MRR) is a phrase that businesses use to describe the amount of money they expect to receive from their subscribers each month
What is a Reverse ETL?
A reverse ETL is the opposite of an ETL, and is used as a development life cycle to test the fitness, quality and readiness of data.
What is a Data Glossary?
A data glossary is a list of terms related to a specific subject, like “data”.
What is a Database?
A database is a tool to collect, store, sort, and manage your data.
What is a Data Warehouse?
A 'data warehouse' is a database managed by an organization.
What is a Snowflake Data Warehouse?
The Snowflake data warehouse is designed to manage big data and allow users to run different kinds of queries on the stored data
What is dbt?
dbt is an open-source command line tool that enables data analysts and engineers to transform data in their warehouse more effectively
What is ETL?
A data movement job can be divided into three steps: Extract, Transform, and Load.
What is a Data Dictionary?
A data dictionary is a central source of information about the data in your organization, business or enterprise
What is A/B Testing?
A/B testing is an experiment where two or more variants of a page are shown to users at random, and statistical analysis is used to determine which variation performs better for a given conversion goal.
Make sense of all your data knowledge in minutes
Try Secoda for free
Book a demo
Made in Canada 🇨🇦
SOC 2 Certified
How to Create a Data Governance Council: A Step-by-Step Guide
Secoda is SOC 2 Certified
What to do as the first data hire at an early-stage startup?
Secoda and Airflow integration
Introducing: Publishing Workflow
Secoda and dbt integration
Data Product Manager