Hi all. I would like to share some issues I’ve been dealing with recently and would like to hear you experiences, ideas and thoughts. Bare with me, this will be slightly longer post.
The issue revolves around usage of LLMs or possibly specialized AI models (if they exist) in validation, compliance enforcing and error correction of Ansible code and other input data. There is a predominant understanding, especially among higher management, that modern AI tools can solve most of the tedious manual human error correction tasks if you just feed it with all of the data and give it instructions on how to “sort this out”.
So here is my example. Let’s say we have around 350 Ansible projects. Projects have a predefined structure of directories for collections, roles, group and host vars, inventory and playbooks. Each project describes one setup consisting of a number of VMs and services deployed to them. There are predefined rules for project and VM naming, required inventory groups, group naming and group hierarchy. We currently rely on human input to correctly define inventory data including VM naming, group membership and other inventory data in general. As it can be expected, we encounter a lot of subtle human made errors, inconsistencies, typos, ordering issues, collisions (two VMs with the same name for example) etc.
Since number of projects are increasing over time and human made errors are piling over time, it is becoming challenging to keep an overview of all of the projects and thousands of VMs and said errors are increasingly becoming a cause of all kind of issues.
That being said, what AI powered tools are out there that could possibly ingest all this data and “sort this out”? Do you have any positive experiences?
My understanding is that for general purpose LLMs, token input limit would be the first obstacle. If I wanted to let LLM only to deal with inventory data, that would be around 1 MB of data (300k tokens roughly). The next issue would be that with this amount od data, LLMs will quickly loose comprehension and start to deviate, make errors itself and hallucinate.