- Knowledge Base
- Setup Data Catalog
- Crawling
-
Roadmap
-
Knowledgebase Documents
-
Installation and System Setup
-
Setup Data Catalog
-
Connectors
-
Data Discovery
-
Self Service
-
Access Management
-
Data Quality
-
Data Literacy
-
Privacy Compliance
-
Reporting
-
Architecture, Security & Releases
-
Developer's Zone
-
Advanced Tools
-
Record of Processing Activities (ROPA)
-
Others
-
Release6.0 User Guide
-
Release6.1 Features
-
Data Catalog
-
News
-
Deactivated_Old
-
Release6.3 Deep Dive Articles
Introduction to OvalEdge Crawling
Crawling is an activity by which OvalEdge collects metadata from various sources. These sources can be databases, data lakes, visualization systems, or reporting systems. In OvalEdge, the crawler records the connection between OvalEdge and third-party databases so that users, with the right permissions, can view the metadata and data .
We have a module in OvalEdge called crawler that connects to a data source, collects, and catalogs all the data elements in the form of metadata that is stored in the OvalEdge data repository.
An index for every stored data element is created which can later be used in data exploration within the OvalEdge Data catalog. OvalEdge crawlers can be scheduled to scan the databases regularly, so they always have an up-to-date index of the data element.
The following is achieved by crawling:
- Connections to source systems
- Retrieval of metadata from source systems
- An ability to understand any changes in the metadata after the first crawl complete.
- The option to dismiss metadata based on regular expression (Regex) programming .