Data Engineering / Compliance AutomationJun 2025
Cloud Data Normalization Pipeline
A data pipeline that takes raw cloud inventory exports and Tenable scan data and turns it into a FedRAMP Appendix M workbook you can actually submit.
Problem
Cloud and security tools each export inventory data in their own shape. FedRAMP wants one specific normalized structure for submission. Somebody has to reconcile that, and doing it by hand is brutal.
Solution
Built a transformation pipeline that maps every inventory export type into one canonical schema, enriches the results with vulnerability data, and writes out a formatted workbook at the end.
Impact
- →Killed the manual reconciliation work across every asset category
- →Produces submission ready Appendix M workbooks straight from the raw exports
- →Turned a tedious compliance reporting task into a repeatable workflow
Architecture
- 01Inventory ZIP contents route by file type
- 02Per format mappers normalize data into one canonical structure
- 03Tenable data joins in by identifier (IP, FQDN, or UUID)
- 04Final output exports into a FedRAMP friendly workbook layout
Capabilities
- ·Multi source CSV normalization
- ·Canonical schema mapping
- ·Tenable enrichment
- ·Formatted XLSX export
Stack
Pythonpandasopenpyxlzipfile