Data Onboarding

Onboard new data sources to Splunk with GDI agent — generates deployment-ready config packages with Magic 8 compliance

Data Onboarding

The GDI Onboarding Agent analyzes your sample logs, matches CIM data models, and generates a complete deployment-ready Splunk app package — including serverclass.conf for deployment server distribution.


The Scenario

You have a new data source to onboard to Splunk. You need inputs, parsing, field extractions, CIM alignment, and a way to deploy configs across forwarders, indexers, and search heads via deployment server. Doing this manually means writing configs by hand, validating Magic 8 compliance, and building Splunk app packages. The GDI agent automates the entire process.

What the Agent Generates

Configuration Files

FilePurpose
inputs.confMonitor stanzas for forwarders — file paths, sourcetype assignment, index routing
props.confIndex-time and search-time parsing — line breaking, timestamp extraction, Magic 8 compliance
transforms.confField extractions, lookups, sed replacements — both index-time and search-time
tags.confCIM tagging — maps events to CIM data models (e.g., Authentication, Network Traffic)
serverclass.confDeployment server classes — host whitelists/blacklists, restart behavior, app-to-host mapping

Multi-App Layout

The agent creates 4 Splunk apps per sourcetype, following best practices for deployment:

AppDeploys ToContains
TA-{sourcetype}_inputsForwardersinputs.conf
TA-{sourcetype}_indexerIndexersprops.conf, transforms.conf (index-time)
TA-{sourcetype}_searchSearch Headsprops.conf, transforms.conf (search-time), tags.conf
TA-{sourcetype}_deploymentDeployment Serverserverclass.conf

This structure separates concerns by Splunk tier and is ready for distribution via deployment server or SH Deployer.

Magic 8 Compliance

Every generated config is validated against the Magic 8 best practices:

  1. Clear sourcetype — unique, descriptive name
  2. Line breaking — handles multi-line events
  3. Timestamp extraction — correct format and timezone
  4. Key field extraction — consistent field names and types
  5. CIM alignment — mapped to relevant data models
  6. Validation — syntax and parsing verification
  7. Documentation — sourcetype ownership and description
  8. Monitoring — ingestion tracking post-deployment

Workflow

  1. Provide sample data — paste or upload a sample of the new data source.
  2. Describe the source — application name, format, expected volume, retention needs.
  3. Agent analyzes — inspects log format, matches CIM data models, identifies fields.
  4. Review generated configs — the agent produces the full multi-app package with all config files.
  5. Validate — the agent runs syntax validation and Magic 8 compliance checks.
  6. Push to GitHub (optional) — create a branch and PR with the generated configs for code review.
  7. Deploy — distribute via deployment server using the generated serverclass.conf.
  8. Monitor — use the Data Quality Check workflow after go-live.

Required Integrations

IntegrationPurpose
Splunk MCPInspect existing configs, validate against live environment
Regex for SplunkGenerate and test field extraction regex
GitHub (optional)Push configs to a repo, create branch and PR for review
Deslicer Observer (optional)Deploy configs via Observer API with approval workflows