Accelerating data modernization and tech debt reduction with AI-enabled delivery

This asset management firm leveraged AI-driven workflows to modernize legacy data systems, reduce technical debt, and accelerate cloud migrations.
About
This privately held investment management firm specializes in a comprehensive suite of investment products, risk management solutions, and advisory services. Managing funds for investors and institutions worldwide, the company operates globally and depends on highly reliable, well-governed data systems to support critical financial decisions.
40-50%
increase in delivery velocity
77%
decrease in total project hours
More
efficient and predictable migration of legacy systems
Challenge
The organization faced significant technical debt across legacy data platforms while undertaking large-scale cloud and analytics migrations.
Core business logic was embedded in:
- Complex SQL stored procedures
- Kedro pipelines
- Legacy C-based services
- Poorly documented workflows that had evolved over time
This created high migration risk and deep dependency on a small group of subject-matter experts. Manual coding, documentation, and validation processes slowed upgrades and increased regression risk.
Approach
Rather than treat modernization as a one-time migration effort, the team restructured delivery itself.
GenAI and agentic workflows were embedded directly into development and migration processes, not just to generate code, but to:
- Reason over existing legacy logic
- Translate SQL procedures and Kedro workflows into dbt models
- Automate documentation and source-to-target mappings
- Generate validation and data quality scripts
- Standardize outputs aligned to governance requirements
By integrating AI into familiar developer environments and CI/CD pipelines, the team accelerated delivery without disrupting client tooling mandates or compliance standards.
Modernization shifted from a manual rewrite project to a scalable, AI-driven way of working.
Solution

GenAI was applied to systematically reduce technical debt and modernize embedded logic across data platforms.
Legacy SQL procedures and Kedro workflows were translated into:
- Modular, maintainable dbt models
- Structured data catalog artifacts
- Automated validation layers
- Version-controlled documentation
In one case, over 1,400 lines of complex SQL were simplified into concise, structured outputs that dramatically improved readability, maintainability, and review efficiency.
A custom Python-based command-line tool leveraging the GitHub Copilot API generated standardized templates (HTML, markdown, YAML, Python) and automated source-to-target mappings. Artifacts were auto-published to Confluence, ensuring documentation remained synchronized with code.
Agentic workflows orchestrated across PyCharm, Python, Jinja templating, and Copilot enabled efficient migrations to AWS and Databricks while optimizing cost, performance, and developer productivity.
Most importantly, validation and standards were embedded directly into delivery workflows — reducing regression risk and increasing confidence in production data.
Key Highlights & Impact
- Automated reduction of complex legacy SQL ~40–50% increase in sprint velocity through workflow standardization and AI acceleration
- A migration effort estimated at ~360 engineering hours delivered in ~84 hours
- Automated translation of legacy SQL and Kedro pipelines into dbt models
- Embedded validation and data quality generation reducing regression risk
- Continuous documentation generated alongside code to eliminate SME bottlenecks
- Predictable, repeatable modernization workflows replacing manual rewrites
- Full migration and decommissioning of legacy C-based services
- Optimized cloud cost and performance through structured migration to AWS and Databricks
Modernization evolved from a slow, manual effort into a scalable delivery model — enabling faster releases, fewer regressions, and greater organizational confidence in data systems.
Platforms
- AWS
- Databricks
- GitHub
- Jinja
- Python
- Spark SQL
- dbt
Capabilities

- Automated code migration across languages, clouds, and data platforms
- AI-assisted generation of validation, testing, and data quality scripts
- Translation of SQL procedures and Kedro workflows into standardized dbt models
- Rapid creation of reusable templates and pipeline artifacts through agentic workflows
- Continuous documentation generation embedded into CI/CD processes
- Scalable pipeline design with support for dynamic enhancements and optimization





