Human Loop CEO Raza Habib shared 5 common errors while building with LLMS

by SkillAiNest

When the Human Loop began in 2020, they were working better to train the latest language models at that time. These models needed manually specified data to do the best. Human Loop’s first product made it easier for anyone to interpret it, while the total amount of manual work decreased rapidly.

But they realized a changing wave.

“About two years ago, we were seeing what’s happening to big language models,” says Humanlope co-founder Raza Habib, and we found that the biggest threat to us as a business was that these biggest models of language would be really good-for example how people would not create AI, and you would not need to interpret it. “

In an ancient move, they began searching for the ax a few months before the start of the chatigon. Instead of helping people interpret their training data, Humanlope will provide teams to assess how their LLM-based AI applications are working well, and will help team members help in collaborating or not.

Raza says, “To lose your company … this is a terrible thing to do.” “So we gave ourselves two weeks of time we would make some fun and go to the people we know (big language models) and see if anyone would pay it. If we could get ten payers in two weeks, it would be a strong indication that it is worth giving importance to the company.”

“Finally, it took us two days,” he says. Today, Human Loop Gusto, Wenta, and Dollingo counts as consumers – working effectively as the LLM playground with their mutual support to find the best indicators, evaluate different models and track changes over time.

Humanlope this week is launching a podcast series named The superior agency (On Apple Pod CastFor, for, for,. SpatifsAnd UTube) Where Raza will talk to others to compare notes in the building at the front of AI to compare what works and what is not in the field of this time. In the first episodes, interviews will be presented with CTOs in Ironcard, Zipare, Source Graph and Hex. All the companies who have made big things with LLM in production – but as Raza has said, “No one is yet an expert, and everyone is learning by doing.”

Keeping this in mind, I asked Raza to break some of the common mistakes they see while building the LLM while making the teams. Here he told me:

Permanent, organized assessment is not in place:

Find out how “good” looks for your AI product output, then find out how to measure it as you build.

Raza says, “If teams have no good way to measure like ‘good’,” he says, “He says,” He will turn his wheel to change things for a long time and do not really know whether they are making any progress. ”

“Everyone wants things that are fast. Everyone wants things that are cheap and accurate. But you are (quality) that are really in terms of specific use for you.”

If you are making an AI chat boot to help someone follow a new language, this means checking output to ensure that only uses words suitable for user skills levels. If you are making an AI coach, it probably means to double check that each of your user’s stated goals is mentioned and addressed.

But there is more to make even more than a few times and make sure it all looks reasonable. The system must be placed in place to check out the output regularly, as the indicator changes and the basic models are ready.

“(Traditional Software Development), You write a piece of code, and whenever you run it, it does the same thing.

He noted, “The biggest mistake of the people is just examples (once).” “This does not give them a strong sense of whether they are improving things or not.”

Don’t pay attention to user feedback (sometimes silent):

“What is ‘good’ that’s a lot!” Notice Reza. “What is a good summary for this call? What is a good sales email? The only right answer is not.”

“What your user says is the final answer,” says Raza. But they do not always call these things out loud.

He noted, “You want to capture different sources of the user’s last impressions.” These can be clear things, such as votes – they lower the small thumb up/thumbs down buttons. But these are also the most common things that users work within your request whether it is well associated with it.

“When you are designing an application, plan ahead to capture the user’s gestures that tell you that if it’s working, you want to design it from the beginning.”

Do not track the quick date closely:

“Another mistake is not treating the immediate management with the same strict management you treat code management,” says Raza.

The indicators you are using will change over time. The key to tracking these changes and knowing why they were made.

“People start doing this and they use (shared documents), they are copying and pasteting things in the silic, and they are losing the date of their experience. New people join the team and it is difficult to know what was tried before. For some months, they will be in production and you will not know what to do.”

Not to fix the model:

Most goals and and when you prove your idea, you may be far away with popular base models. But finally, Raza advised, you would like to fix them for your needs. Good fine toning will reduce you better results, low delays and long -term costs.

“We advise everyone that quick engineering should start, because it is the easiest, fastest and most powerful thing,” says Raza. “But if you fix your models later, you can get an order off -magnetic cost savings.”

He noted, “The best way to think about fine toning is as a correction.” “You want to avoid premature correction, but once you confirm that your product is demanding, then it should be the focus.”

Domain experts do not write hints:

If you are creating LLM products for a particular vertical or industry, bring people who really are Know Title to help write indicators and assess the output – don’t rely on engineers to do alone. Large models of language, clearly, are about the language. Language is given importance, and the dictionary of different industries is deep.

“This is the work that is best done by domain experts,” says Raza. “This is one of the things that is clear in the retreat, but in the beginning it was not clear.”


If you are building with LLM, be sure Check the Human Loop hereAnd find a new pod cast of Raza for AI builders, high authority, Here on YouTube.

You may also like

Leave a Comment

At Skillainest, we believe the future belongs to those who embrace AI, upgrade their skills, and stay ahead of the curve.

Get latest news

Subscribe my Newsletter for new blog posts, tips & new photos. Let's stay updated!

@2025 Skillainest.Designed and Developed by Pro