- Cerebral Valley
- Posts
- Refuel is rethinking data-management in the age of AI đ
Refuel is rethinking data-management in the age of AI đ
Plus: Rishabh & Nihit on Refuel LLM, data enrichment and AI research...
CV Deep Dive
Today, weâre talking with Rishabh Bhargava and Nihit Desai, Co-Founders of Refuel.
Refuel is a platform that helps enterprise ML teams clean, label and enrich data at scale. The companyâs mission is to use AI to power data workflows such as data labeling, cleaning, enrichment and extraction, so that ML and product teams across startups and enterprises can build data-driven products and train their own AI models with confidence. Founded in 2021, Refuel is built upon the idea that data is the true lifeblood of the current AI revolution, and that without proper data management processes in place, AI will not reach its full potential in the enterprise stack.
Today, Refuel has hundreds of teams using its platform for data cleaning, enrichment and extraction, including at companies like Enigma, Spot AI and TeachFX. In 2023, the young startup announced a $5m Seed round led by General Catalyst and XYZ Ventures, in order to expand its team and work towards a public launch.
Thrilled to announce the launch of @RefuelAI! We are building a platform to create, enrich and label datasets with Large Language Models (LLMs). More details in the đ§ľ
â Nihit Desai (@nihit_desai)
10:07 PM ⢠Jun 15, 2023
In this conversation, Rishabh and Nihit walk us through the founding premise of Refuel, why AI data is the lifeblood of todayâs AI products, and Refuelâs goals for the next 12 months.
Letâs dive in âĄď¸
Read time: 10 mins
Our Chat with Rishabh and Nihit đŹ
Rishabh and Nihit - welcome to Cerebral Valley! Firstly, tell us about your background and what led you to co-found Refuel?
Hey there! Iâm Rishabh, one of the co-founders of Refuel. My co-founder Nihit and I both come from machine learning and data systems backgrounds, and we met during grad school at Stanford. Weâve known each other for about a decade now, and founded Refuel just over two years ago.
Before Refuel, I was an early ML engineer at a company called Primer.AI in 2017, where we were building a tool to ingest the world's unstructured text information - things like news, social media, SEC filings - and give people an interface where they could just type a question and get a full-fledged report to read. So, very much ahead of our time back then - this now takes a few lines of code! I then had a stint at a company building data pipelines at the terabytes scale, and that company got acquired by Cloudera a couple of years ago.
My co-founder, Nihit, was at Meta working on trust and safety and content moderation problems prior to Refuel. He was focused on building scalable techniques for understanding large swaths of multimodal data, and then scaling it out to Facebook scale in near real time. He left Facebook at the end of 2021, and weâve both been building Refuel ever since.
We started Refuel to accelerate the era of AI abundance by solving the core bottleneck â data. No company has ever told us that theyâve got plenty of data thatâs clean, structured and ready to use. Today, most businesses are generating and ingesting tons of data that remains unused simply because it's unstructured and hard to deal with. At the same time, there is a realization that data is the lifeblood of modern products - we're generating and collecting data at exponential scale, but our ability to understand it and make it usable is bottlenecked by human effort.
It doesnât have to be this way - our goal with Refuel is to unlock the value of AI for every business.
Label, clean and enrich text datasets with Large Language Models (LLMs)
github.com/refuel-ai/autoâŚ
â Alex Yanko đşđŚ (@LeopolisDream)
8:35 PM ⢠Jun 20, 2023
For an ML engineer or team who hasnât heard of Refuel before, how would you describe what Refuel does?
Refuel is a platform that helps teams clean, label and enrich data at scale by leveraging LLMs, while delivering superhuman (and better than GPT-4) accuracy.
If youâve ever been in a room where someone has said - âWeâd love to start this project, but we don't have engineering resources to get our data in order for the next two quartersâ or, âif weâre to go out and clean and label our data, we'd have to hire this massive team of contractors, and who's going to manage that?â - then Refuel is meant for you. Data is a massive chore for most teams and if the problem of getting to clean, structured data doesn't get solved, then a lot of the promise of AI will simply never be realized.
Working with Refuel is a simple three-step process. First, point Refuel at where your data sits and describe your data operation in natural language. Refuelâs models will run on your data to produce the desired outputs, while flagging any low-confidence results for review. Second, deploy this task to run in batch or real-time mode (at any latency), by training and deploying custom models fine-tuned for your tasks. Finally, set up your task for continual improvement with further feedback and retraining built-in.
All in all, Refuel is a complete solution for teams to solve data cleaning, data extraction and data labeling problems at scale.
Who are your users today? Whoâs finding the most value in using Refuel?
Our customers are startups and enterprises across several different verticals - such as financial services, marketplaces and e-commerce, where data quality, classification, extraction and matching problems run rampant. Our users include data scientists who want to create labeled datasets quickly, developers who are building highly accurate LLM applications, and even operations teams looking to automate time-consuming workflows.
As a concrete example: imagine you're a marketplace and youâre ingesting products from different vendors, improving their quality/descriptions, and then making these products available on your platform for end consumers. A common pain point you'll run into is standardizing this data into your schema and then improving the quality to make the product look appealing to consumers. So, youâll either spend a bunch of human time reviewing and editing data, or you'll have to hire a large machine learning team for training and improving many different ML models systems.
Instead, you can leverage Refuel by pointing us at the raw datasets being ingested, and describing how you want your data to be normalized and enriched. You can write down simple instructions or use one of our templates - for example, an enrichment template that we have - and then Refuel will help you pick the best LLMs for your task that you can further customize with simple thumbs-up and thumbs-down feedback.
The value for teams is a simple, consistent workflow for leveraging, fine-tuning and deploying high quality models and applications to production â typically outperforming GPT-4 and human quality, while scaling to any data volumes and with minimal effort.
Refuel is super cool. Theyâre able to address the issue of data labeling at scale, itâs truly impressive. They turn the process into âuse Refuel, focus on the data science.â @RefuelAI#MachineLearning#LLMs
â Rohith Desikan (@RohithDesikan)
4:03 PM ⢠Feb 1, 2024
It feels like youâre taking a full-stack approach to the data enrichment problem. Would you say thatâs a unique differentiation point between yourselves and competitors?
This is a huge part of why we're building Refuel the way we're building it â as an end-to-end platform. Many of our users are people who are not machine-learning experts - these are folks who are experts in their domain and know how they want the problem solved, but don't have the ML expertise themselves. There are all sorts of questions that come up when leveraging LLMs to solve data problems â âWhat does a good prompt look likeâ, âWhich LLMs are good for my task and how do I evaluate them?â, âWhen should I use few-shot prompting and which examples should I use?â âHow do I prepare datasets for training?â, and more. But, as long as you can point Refuel at your data, and you know what the correct output should look like, you can get tons of leverage from Refuel.
Essentially, Refuel ends up becoming a way for users to turn their internal expertise into an operationalized process that produces LLMs that are going to work well in production. Itâs highly valuable having all of this in one place, such that the product can suggest how to structure your prompts and improve them behind the scenes, all the way to helping you assess the quality of models/data for fine-tuning and continually improving in production. Refuel ends up being one consolidated place where a lot of this can happen.
Additionally, we also have our own LLM that we've instruction-tuned on close to 3,000 data sets and many billions of tokens across multiple verticals, which is purpose-built for what we think of as data applications - classification, enrichment, extraction and more. What weâve found with customers is that performance out of the gate is oftentimes at parity or above the most powerful LLMs like GPT-4. More importantly, it's very data-efficient to fine-tune it for specific customer applications - with 500 examples and less than 30 minutes of training, you can get to something that beats a fine-tuned GPT 3.5 or GPT-4.
Weâre thrilled to announce Refuel LLM, a large language model built for data labeling and enrichment tasks. Outperforms trained human annotators, GPT-3.5-turbo, PaLM-2, Claude and a host of open-source LLMs on a benchmark of data labeling tasks: refuel.ai/blog-posts/annâŚ
Further,⌠twitter.com/i/web/status/1âŚ
â Nihit Desai (@nihit_desai)
7:27 PM ⢠Oct 17, 2023
Walk us through the customer experience of using Refuel - are you seeing any use-cases that are extra-compelling?
We can illustrate this through one of our customers, where their internal revenue operations team are our customers. Oftentimes, this team will want higher quality leads and to hyper-personalize their messaging when talking to a potential customer, and so theyâll leverage Refuel for enriching potential leads - information like the companyâs size, location of headquarters, industry and so on. One entertaining story is that theyâll ask Refuel to generate a list of fun history facts about their customerâs location before they go and talk to them - which is a use case we didnât anticipate when we started working with them. Once theyâve set this up within Refuel, it runs automatically for every new data lead that enters their pipeline.
Another trend we're seeing is that a lot of customers are interested in Refuel because they're already thinking about fine-tuning and customizing models for specific use-cases. Natural questions that come to their mind are âHow far can we take this? Can we customize models on a per industry level, and customize LLMs on a per-customer basis?â This is one of the trends weâre finding super interesting because, to be able to do this successfully, you need very fine-grained control over your data and the set of models that you're training, and have the relevant infrastructure to be able to pull this off. If done well, the level of quality that you can achieve for any given customer is really high.
Whatâs the biggest technical challenge associated with building Refuel?
There are multiple, but the foremost is around LLM output quality and reliability. LLMs are fundamentally a very new piece of technology, especially in terms of leveraging them at scale and applying them to solve real-world problems. Even though the research has been around for five or six years, the community collectively has twelve months of experience, and the bar for reliability and quality is fairly high for most customers that we work with across critical business use cases. Specifically, the challenge is getting high-quality outputs that you can then rely on as a business.
To give you a concrete example, one of our customers has a lot of domain-specific data for which they've built a large list of rules and heuristics over a decade. They've had a large team of expert humans building and managing this list, which is complex as a single change may break a bunch of things elsewhere. Now, this customer is using Refuel to replace a lot of these rules with a single LLM that is exceptional at that one task - and we're seeing success there. But, being able to deliver that experience for all types of business problems that exist - across data classification, extraction and more - and making it turnkey, is one of the hardest parts of building Refuel.
At @RefuelAI, we set out to evaluate the performance of LLMs like GPT-4, PaLM-2 and open source models for autolabeling datasets across a range of NLP tasks. Excited to share our learnings so far, and next steps with the community in this đ§ľ
â Nihit Desai (@nihit_desai)
5:02 PM ⢠Jun 16, 2023
How do you incorporate AI research into your product development process, given the pace of breakthroughs on a weekly basis?
Firstly, Cerebral Valley has been a great source of information for keeping up with new AI developments - almost as a proxy for whatâs top of mind and important for the community. Itâs true that every week, a bunch of new ideas come up and we leverage Twitter/X to stay on top of the best research.
In terms of actually deciding what research to incorporate into our product development roadmap, we donât have a very systematic process today â a lot of it is building intuition for how relevant a piece of research will be to the product that we're building and to our customer base. We try to have a sense of the impact of new research before going too deep into it, and our priorities are driven by customer conversations, or by our aggressive product roadmap. As an example, very early on we heard from users that estimating âconfidenceâ of LLM outputs to better detect hallucinations was critical, and so we spent a number of cycles going deep into the research.
Labeling LLM data with confidence is hard.
This post by @nihit_desai and @BansalDhruva from @RefuelAI was surprisingly valuable for how we're thinking about these things in @cardinalappio đâ Wiz (@nadavwiz)
8:52 PM ⢠Feb 1, 2024
What are you going to be most focussed on for the next 6-12 months? Customer growth? Infrastructure? ML research?
All of the above. Weâve seen significant growth in terms of the number of customers using Refuel, and we expect to continue to grow that significantly. For example, we've gone from having processed a million data points across our customers a few months ago, to 12 billion data points today. Thatâs a 1,000x increase in the volume of data that weâve processed - and so weâre seeing that scale happen in real time.
You either die a cto or live long enough to do customer support. Acc or â ď¸
â Nihit Desai (@nihit_desai)
5:20 AM ⢠Feb 2, 2024
Additionally, a lot of our effort over the next 12 months is going to be spent on improving our platform - making it more robust, capable and friendlier for teams to get to their desired outcomes. So, for us, the next 6-12 months are very much about growing the number of customers that we have, while making sure that our platform continues to become better, more robust, and able to handle more use cases.
Lastly, tell us about the team culture at Refuel. What do you look for in prospective team members, and are you hiring?.
Culture at Refuel can be framed along a few different axes. When hiring, we look for a demonstration of real-world impact, and the desire to create more. Weâre also extremely customer-centric â one of the ways this shows up for our engineering team is that weâll happily leverage everything that's been built already, because ultimately, what we care about most is delighting our customers.
We also place a heavy premium on moving fast, given how small of a team we are and how fast the ecosystem is moving. We're okay with shipping things that might be less than perfect as long as we maintain this culture of learning, iterating, and shipping the next version out as soon as possible. Tied to that, we also really look for an âownership mentalityâ around the work - this is critical for us given our broad product surface area, and the number of important decisions we need to make on behalf of our customers every single day.
Conclusion
Read our past few Deep Dives below:
If you would like us to âDeep Diveâ a founder, team or product launch, please reply to this email ([email protected]) or DM us on Twitter or LinkedIn.