On the 15th of October Erik MD gave a quick demo at AI Tinkerers Paris about our first version of BigWig. This is a short writeup based on these slides Agents for building ML models.
The slides and the demo are a bit outdated, as we are moving fast, but the core ideas are the same.
There are three parts to the talk/demo:
The Infer platform splits into 3 layers stacked on top of each other:
An unified real-time SQL interface combining data retrieval, statistical modelling and large-scale compute in one single and simple interface.
A user interface that abstracts away the core layer, enabling users to build, deploy and integrate machine learning models across their data stack.
An agentic framework built on top of the core layer to enable autonomous workflows and analysis.
The platform consists of a number of interconnected modules. All of which the agents operating on top of the Agent Layer can use and interact with. These modules allow the agents to perform a wide range of tasks, from data retrieval and simple EDA analysis to model building and large scale inference.
The common interface for interacting with the platform is through SQL-inf.
SQL-inf is a unified interface for
As an example of how SQL-inf works, this is how easy it is to model, build and execute a custom lead scoring model on your HubSpot data in a single, simple SQL statement:
SQL-inf is Perfect for Agents!
SQL-inf and the Core layer lends itself extremely well to agentic flows by giving them a simple layer of software abstraction for querying data and building machine learning and statistical models.
LLMs are great at acting as a good analyst at a high-level - they have read a lot of medium articles and course note about that!
They are also pretty decent at writing SQL - maybe not so much zero-short text->SQL generation but through an iterative process, with a clear measure of quality for each step, they tend to converge to a good, stable answer
They are also really good at tool use - if the tools are well-defined, behave well and return well-structured outputs
We have well-defined software abstractions that wrap complicated implementations - reducing variability and instability
We have an objective scoring model for each step, in our cases statistical measures on models combined with some heuristics
This scoring allow the agents to effectively search for a converging solution through chained actions - locally unstable, but globally stable
We use LLMs on well-learned domains: high-level analyst reasoning(good) and SQL generation(okay-ish)
A simple example of an agentic flow for building a machine learning model
Higher level reasoning agents - using current single flow agents in a multi-agent environment to solve higher level task
Reinforcement learning on reasoning steps - new reasoning model on top of current agents, to improve local stability of reasoning chains
Is this something akin to System 2 thinking?
I don't know - but it is definitely pretty impressive and it is well beyond System 1 smarter RPA/"workflow automation" thinking - perhaps System 1.5?
If not quite System 2 yet, I think we have the beginning of a general autonomous system for performing statistical modelling and analysis - and that is definitely, from human perspective, System 2!
It also brings up some new interesting aspects for the future of analays: