The Data and Analytics team at G-Research provides core infrastructure (databases, streaming, batch processing and workflow systems) used across research, cybersecurity and risk management.
Like most of our platform teams, we primarily use Python. While we’ve had access to AI coding assistants for some time, we hadn’t yet built an LLM-powered system ourselves.
This project changed that.
We developed a team coding standards review tool that checks git diffs against our internal Python coding standards using an LLM. It runs in CI/CD on every feature branch and posts its findings as a PR comment, so engineers can see standards violations before the review process begins.
It began as a 10%-time project, aligned with a broader Data and Analytics Engineering north star focused on delivery, reliability, and reuse.
Moving from a “tool that kind of works” to “a tool you’d trust in CI” involved solving issues that don’t exist in traditional software. This post shares what we learned along the way.