Skip to main content
Version: Next

OpenAPI

Incubating

Important Capabilities

CapabilityStatusNotes
DescriptionsExtracts endpoint descriptions and summaries from OpenAPI specifications.
DomainsDoes not currently support domain assignment.
Extract OwnershipDoes not currently support extracting ownership.
Extract TagsExtracts tags from OpenAPI specifications.
Schema MetadataExtracts schemas from OpenAPI specifications for GET, POST, PUT, and PATCH methods.

OpenAPI source implementation for DataHub ingestion.

This class provides the concrete implementation of the OpenAPI source, configured specifically for OpenAPI specifications.

CLI based Ingestion

Starter Recipe

Check out the following recipe to get started with ingestion! See below for full configuration options.

For general pointers on writing and running a recipe, see our main recipe guide.

source:
type: openapi
config:
name: my_api
url: https://api.example.com
swagger_file: openapi.json

sink:
type: "datahub-rest"
config:
server: 'http://localhost:8080'

Config Details

Note that a . is used to denote nested fields in the YAML recipe.

FieldDescription
name 
string
Name of ingestion.
swagger_file 
string
Route for access to the swagger file. e.g. openapi.json
url 
string
Endpoint URL. e.g. https://example.com
bearer_token
One of string, null
Bearer token for endpoint authentication.
Default: None
enable_api_calls_for_schema_extraction
boolean
If True, will make live GET API calls to extract schemas when OpenAPI spec extraction fails. Requires credentials (username/password, token, or bearer_token). Only applicable for GET methods.
Default: True
forced_examples
object
If no example is provided for a route, it is possible to create one using forced_example.
Default: {}
get_token
object
Retrieving a token from the endpoint.
Default: {}
password
string
Password used for basic HTTP authentication.
Default:
proxies
One of object, null
Eg. {'http': 'http://10.10.1.10:3128', 'https': 'http://10.10.1.10:1080'}.If authentication is required, add it to the proxy url directly e.g. http://user:pass@10.10.1.10:3128/.
Default: None
schema_resolution_max_depth
integer
Maximum recursion depth for resolving schema references. Prevents infinite recursion from deeply nested or circular references. Default is 10 levels.
Default: 10
token
One of string, null
Token for endpoint authentication.
Default: None
username
string
Username used for basic HTTP authentication.
Default:
verify_ssl
boolean
Enable SSL certificate verification
Default: True
ignore_endpoints
array
List of endpoints to ignore during ingestion.
Default: []
ignore_endpoints.object
object

This connector ingests OpenAPI (Swagger) API endpoint metadata into DataHub. It extracts API endpoints from OpenAPI v2 (Swagger) and v3 specifications and represents them as datasets in DataHub, allowing you to catalog and discover your API endpoints alongside your data assets.

Concept Mapping

This ingestion source maps the following Source System Concepts to DataHub Concepts:

Source ConceptDataHub ConceptNotes
"OpenAPI"Data Platform
API EndpointDatasetSubtype API_ENDPOINT

Capabilities

The OpenAPI source extracts metadata from OpenAPI specifications and optionally makes live API calls to gather schema information. It supports:

  • OpenAPI v2 (Swagger) and v3 specifications - Automatically detects and processes both formats
  • Schema extraction from specifications - Extracts schemas directly from OpenAPI spec definitions (preferred method)
  • Schema extraction from API calls - Optionally makes GET requests to endpoints when schema isn't in spec (requires credentials)
  • Multiple HTTP methods - Supports GET, POST, PUT, and PATCH methods with 200 response codes
  • Browse path organization - Endpoints are organized by their path structure in DataHub's browse interface
  • Tag extraction - Preserves tags from OpenAPI specifications
  • Description extraction - Extracts endpoint descriptions and summaries

Schema Extraction Behavior

The source uses a multi-step approach to extract schemas for API endpoints:

  1. OpenAPI Specification (Primary) - The source first attempts to extract schemas directly from the OpenAPI specification file. This includes:

    • Response schemas from 200 responses
    • Request body schemas for POST/PUT/PATCH methods
    • Parameter schemas when available
  2. Example Data (Secondary) - If schemas aren't fully defined, the source looks for example data in the specification

  3. Live API Calls (Optional) - If enable_api_calls_for_schema_extraction=True and credentials are provided, the source will make GET requests to endpoints when:

    • Schema extraction from the spec fails
    • The endpoint uses the GET method
    • Valid credentials are available (username/password, token, or bearer_token)
note

API calls are only made for GET methods. POST, PUT, and PATCH methods rely solely on schema definitions in the OpenAPI specification.

tip

Most schemas are extracted from the OpenAPI specification itself. API calls are primarily used as a fallback when the specification is incomplete.

Schema Extraction Priority

When multiple HTTP methods are available for an endpoint, the source prioritizes extracting metadat from methods in this order:

  1. GET
  2. POST
  3. PUT
  4. PATCH

The description, tags, and schema metadata all come from the same priority method to ensure consistency.

Browse Paths

All ingested endpoints are organized in DataHub's browse interface using browse paths based on their endpoint path structure. This makes it easy to navigate and discover related endpoints.

For example:

  • /pet/findByStatus appears under the pet browse path
  • /pet/{petId} appears under the pet browse path
  • /store/order/{orderId} appears under storeorder

Endpoints are grouped by their path segments, making it easy to find all endpoints related to a particular resource or feature.

Prerequisites

OpenAPI Specification Access

  • The OpenAPI specification file must be accessible via HTTP/HTTPS
  • The specification should be in JSON or YAML format
  • OpenAPI v2 (Swagger 2.0) and v3.x specifications are supported

Authentication (for API calls)

If you want to enable live API calls for schema extraction (enable_api_calls_for_schema_extraction=True), you'll need to provide authentication credentials. The source supports:

  • Bearer token authentication
  • Custom token authentication
  • Dynamic token retrieval (GET or POST)
  • Basic authentication (username/password)
note

Authentication is only required if you want to enable live API calls. Schema extraction from the OpenAPI specification itself does not require authentication.

Authentication Methods

Bearer Token

source:
type: openapi
config:
name: my_api
url: https://api.example.com
swagger_file: openapi.json
bearer_token: "your-bearer-token-here"

Custom Token

source:
type: openapi
config:
name: my_api
url: https://api.example.com
swagger_file: openapi.json
token: "your-token-here"

Basic Authentication

source:
type: openapi
config:
name: my_api
url: https://api.example.com
swagger_file: openapi.json
username: your_username
password: your_password

Dynamic Token Retrieval

The source can retrieve a token dynamically by making a request to a token endpoint. This is useful when tokens expire and need to be refreshed.

source:
type: openapi
config:
name: my_api
url: https://api.example.com
swagger_file: openapi.json
get_token:
request_type: get # or "post"
url_complement: api/auth/login?username={username}&password={password}
username: your_username
password: your_password
note

When using get_token with request_type: get, the username and password are sent in the URL query parameters, which is less secure. Use request_type: post when possible.

Forced Examples

For endpoints with path parameters where the source cannot automatically determine example values, you can provide them manually using forced_examples:

source:
type: openapi
config:
name: petstore_api
url: https://petstore.swagger.io
swagger_file: /v2/swagger.json
forced_examples:
/pet/{petId}: [1]
/store/order/{orderId}: [1]
/user/{username}: ["user1"]

The source will use these values to construct URLs for API calls when needed.

Ignoring Endpoints

You can exclude specific endpoints from ingestion:

source:
type: openapi
config:
name: my_api
url: https://api.example.com
swagger_file: openapi.json
ignore_endpoints:
- /health
- /metrics
- /internal/debug

Examples

Basic Configuration (Schema from Spec Only)

source:
type: openapi
config:
name: petstore_api
url: https://petstore.swagger.io
swagger_file: /v2/swagger.json
enable_api_calls_for_schema_extraction: false

sink:
type: "datahub-rest"
config:
server: "http://localhost:8080"

With API Calls Enabled

source:
type: openapi
config:
name: petstore_api
url: https://petstore.swagger.io
swagger_file: /v2/swagger.json
bearer_token: "${BEARER_TOKEN}"
enable_api_calls_for_schema_extraction: true

sink:
type: "datahub-rest"
config:
server: "http://localhost:8080"

Complete Example with All Options

source:
type: openapi
config:
name: petstore_api
url: https://petstore.swagger.io
swagger_file: /v2/swagger.json

# Authentication
bearer_token: "${BEARER_TOKEN}"

# Optional: Enable/disable API calls
enable_api_calls_for_schema_extraction: true

# Optional: Ignore specific endpoints
ignore_endpoints:
- /user/logout

# Optional: Provide example values for parameterized endpoints
forced_examples:
/pet/{petId}: [1]
/store/order/{orderId}: [1]
/user/{username}: ["user1"]

# Optional: Proxy configuration
proxies:
http: "http://proxy.example.com:8080"
https: "https://proxy.example.com:8080"

# Optional: SSL verification
verify_ssl: true

sink:
type: "datahub-rest"
config:
server: "http://localhost:8080"

Limitations

  • API calls are GET-only: Live API calls for schema extraction are only made for GET methods. POST, PUT, and PATCH methods rely solely on schema definitions in the OpenAPI specification.
  • Authentication required for API calls: If enable_api_calls_for_schema_extraction=True, valid credentials must be provided.
  • 200 response codes only: Only endpoints with 200 response codes are ingested.
  • Schema extraction from spec is preferred: The source prioritizes extracting schemas from the OpenAPI specification. API calls are used as a fallback.

Troubleshooting

No schemas extracted

If schemas aren't being extracted:

  1. Check the OpenAPI specification - Ensure your spec includes schema definitions in responses or request bodies
  2. Enable API calls - Set enable_api_calls_for_schema_extraction: true and provide credentials
  3. Check authentication - Verify your credentials are correct if API calls are enabled
  4. Review warnings - Check the ingestion report for warnings about specific endpoints

Endpoints not appearing

If endpoints aren't appearing in DataHub:

  1. Check ignore_endpoints - Ensure endpoints aren't in the ignore list
  2. Verify response codes - Only endpoints with 200 response codes are ingested
  3. Check OpenAPI spec format - Ensure the specification is valid OpenAPI v2 or v3

Authentication errors

If you see authentication errors:

  1. Verify credentials - Check that username/password or tokens are correct
  2. Check token format - Bearer tokens should not include the "Bearer " prefix
  3. Review get_token configuration - Ensure the token endpoint URL and method are correct

Code Coordinates

  • Class Name: datahub.ingestion.source.openapi.OpenApiSource
  • Browse on GitHub

Questions

If you've got any questions on configuring ingestion for OpenAPI, feel free to ping us on our Slack.