SQL Server Connector

SQL Server with SQL Server Authentication

Connect to SQL Server

Crawling is a process of collecting information about data from various data sources like on-premise and cloud databases, Hadoop, visualization software, and file systems. When an OvalEdge crawler connects to a data source, it collects and catalogs all the data elements (i.e. metadata) and stores it in the OvalEdge data repository. The crawler creates an index for every stored data element, which can later be used in data exploration within the OvalEdge Data catalog which is a smart search. OvalEdge crawlers can be scheduled to scan the databases regularly, so they always have an up-to-date index of the data element.

Data Sources

OvalEdge crawler integrates with various data sources to help the users to extract metadata and build a data catalog. In this document, you can see how to make a connection to your SQL server instance and crawl the databases and schemas.

Connect to the Data

Before you can crawl and build a data catalog, you must first connect to your data. OvalEdge requires users to configure a separate connection for each type of data source. The users must enter the source credentials and database information for each type of connectivity. Once a data connection is made, a simple click of the Crawl button starts the crawling process.

You can use either SQL Server Authentication/Windows Authentication to connect to your SQL server instance.

Configure a new database connection

  1. Select the Administration tab and click on it to expand the functions.
  2. Select the Crawler.
  3. Click the New Connection from the Crawler window. The dialog box for configuring a database connection is displayed.
  4. Search for SQL Server in the Search connector.
  5. Enter the connection details.
  6. Enter the Driver name and Connection string manually.
    Alternatively, set the Connection string toggle button to automatic, to get the details automatically from the credentials provided.

  7. Click Validate to test the database connection.

  8. When validation completes, click Save to finish the configuration of the database connection. 
    A dialog box that the database added successfully is displayed.

Connection details


Default/Sample value*


Database Type

SQL Server

Select the database type from the list


SQL Server Authentication

This is the default authentication

(Only for SQL server connection)

Connection Name

Requires input

Reference name for database identification in OvalEdge

Hostname / IP Address

Requires input

Database instance URL(on-premises/cloud-based) **

Port Number

Requires input

Default 1433

Port number of the database instance


Requires input

A unique name that identifies the selected database

(Only for SQL Server connection)

Driver Name


JDBC driver name for SQL server

Connection String


SQL server

*- Requires Client-specific input

Crawl a database connection

In OvalEdge, each data source can be crawled at a Schema level. In other words, when there are multiple schemas within a database, you have an option to select a specific schema and crawl. 

To crawl a specific Schema or multiple schemas,

  1. Select the Administration tab and click on it to expand the functions.
  2. Select the Crawler. 
  3. Select a database from the list.
  4. Click the Crawl/Profile button on the crawler page.
    The dialog box to select Important schema for crawling and Profiling is displayed.
  5. Click the Checkbox next to the schema name and select a single schema or multiple schemas to crawl and profile.
  6. To crawl all schemas, click Select All.
  7. Click OK to initiate Crawling the database connection immediately.

    Alternatively, select Advance options to schedule a  Crawl and Profile action 


  8. A dialog box displays that a crawl job is Initiated.
  9. Select the Jobs tab in the object explorer(left side of the window) to know the status of the job submitted.
  10. Check the Job step status and click the  Refresh button.

    INIT -                            Advanced Job is initiated

    RUNNING -                  Job is being processed

    WAITING  -                  Submitted Job is on Queue to be processed

    SUCCESS  -                  Job is completed successfully

    ERROR -                      Job has failed. Check error message in the Logs

    KILLED -                     The initiated job has been killed.

    PARTIAL SUCCESS - The initiated job has completed partially.
  11. Go to the Crawl window to check the Last Crawl Status of the selected Schemas.