Features | Architecture | Documentation References | Compatibility| Log Ingestion Examples | Feedback | Legal Information
The Splunk Integration project is a non-supported bidirectional connector consisting of three main components as depicted in the architecture diagram:
We also provided extensive documentation for Log Collection to ingest, store, and process logs on economical and performant Delta lake.
Run Databricks SQL queries right from the Splunk search bar and see the results in Splunk UI
Execute actions in Databricks, such as notebook runs and jobs, from Splunk
Use Splunk SQL database extension to integrate Databricks information with Splunk queries and reports
Push events, summary, alerts to Splunk from Databricks
Databricks Add-on for Splunk, notebooks and documentation provided in this project are compatible with:
This project also provides documentation and notebooks to showcase specifics on how to use Databricks for collecting various logs (a comprehensive list is provided below) via stream ingest and batch-ingest using Databricks autoloader and Spark streaming into cloud Data lakes for durable storage on S3. The included documentation and notebooks also provide methods and code details for each log type: parsing, schematizing, ETL/Aggregation, and storing in Delta format to make them available for analytics.
Data collection sources with notebooks and documentation are included for the following sources:
Issues with the application? Found a bug? Have a great idea for an addition? Feel free to file an issue or submit a pull request.
This software is provided as-is and is not officially supported by Databricks through customer technical support channels. Support, questions, help, and feature requests can be communicated via email -> cybersecurity@databricks.com or through the Issues page of this repo.