Heading

1

What’s a Rich Text element?

The rich text element allows you to create and format headings, paragraphs, blockquotes, images, and video all in one place instead of having to add and format them individually. Just double-click and easily create content.

Static and dynamic content editing

A rich text element can be used with static or dynamic content. For static content, just drop it into any page and begin editing. For dynamic content, add a rich text field to any collection and then connect a rich text element to that field in the settings panel. Voila!

How to customize formatting for each rich text

Headings, paragraphs, blockquotes, figures, images, and figure captions can all be styled after a class is added to the rich text element using the "When inside of" nested selector system.

Resource

Key Elements for Timely Completion of Machine Learning Projects

Key Elements for Timely Completion of Machine Learning Projects
No items found.
Whether you’re looking into AI for the first time or looking to refine your approach, this guide is packed with easy-to-follow insights and advice. From project planning to tackling the unexpected, we’ll walk you through everything you need to know to keep things running smoothly. Ready to make your AI project a success? Let’s jump in!
Download it now.

Download "Key Elements for Timely Completion of Machine Learning Projects" to Learn

  • How to establish clear project milestones and deadlines.
  • The best practices for gathering and preparing data efficiently.
  • Tips for fostering collaboration between data scientists and developers.
  • Strategies for managing unexpected challenges and staying on track.
  • Tools to streamline deployment and ensure project success.

Understanding the Challenges of AI Projects

It’s difficult to find a modern organization that isn’t taking a hard look at artificial intelligence. Not only is it all over the news, but new advancements are unlocking capabilities that weren’t possible even five years ago. It was unthinkable to outsource customer support to AI in 2020 - now, it’s becoming the default.

Going from idea to production isn’t nearly as straightforward as, say, a software engineering project. This is uncharted territory for most companies, and few have the expertise to handle it themselves. AI solutions sit at the crossroads of engineering and data science - bringing with them, the problems of each. 

There’s a lot to consider - building out a mature in-house data science practice is not easy. 

  • How would you even interview the FIRST ML engineer to start building a team? 
  • Do you know what to ask? 
  • Will you understand the response? To an outsider, it almost sounds like a foreign language.
  • Hiring just one engineer puts you at risk of losing all knowledge when they leave. Even if you successfully hire more, do you have enough AI challenges and opportunities to keep them interested for years to come?

Instead of spending years figuring all that out, a vetted team of experts can get you to a solution quicker. AI agencies (like ours, NineTwoThree) bring the experience, you bring the problem and the data.

It’s important to understand what you’re signing up for, though. 

If you’re at the point where you’re considering an ML project for your company, then congratulations - you’re getting in at the right time. This technology is going to change the ways a lot of people interact with the world. But if there are two words to sum up the AI hype, it’s “unrealistic expectations.” 

There are so many considerations at every stage of the process - data, model selection, training, fine-tuning, testing, bringing to production…each stage brings complications and delays. And if you’re partnering with an agency, you’re going to want to make sure things are on-time - otherwise, costs can add up, and projects can sit idle.

Let’s go over each stage of the process, and understand what it takes to deliver on-time.

Understand the knowledge gaps

It truly takes a village to make a world-class AI solution. 

You’ll need expertise across data science, software engineering, DevOps, legal, and product…and that’s if the app DOESN’T face any customers.

A few common scenarios:

  1. Is there a robust data science practice at your company, but no MLOps and DevOps specialists to productionize the data insights? 
  2. Are there engineers who build scalable systems, but don’t have any data science expertise? 
  3. Is the marketing team collecting granular, quality data, with nowhere to send it?

Understanding these gaps is key, and step one.

“If you’re building an AI solution for the restaurant industry for example, it’s important everyone has a base level of domain knowledge. What features are important for restaurants? What influences their revenue? Getting this information early can save time later.” says Vitalijus Cernej, a Machine Learning Engineer at NineTwoThree.

Trust us, the sentence “we’ll handle it ourselves” should only come up if you’re CERTAIN the expertise is there.

Agree on a (realistic) timeline

One is reminded of Hofstadter’s Law:

“It always takes longer than you expect, even when you take into account Hofstadter's Law.”

Any project is going to have unpredictability in the timeline, but that doesn’t mean there shouldn’t be expectations.

“Experience is the cornerstone of efficiency. The real question isn't if a problem will arise, but when.” says Vitalijus.

This is where leaning on the experts can help. They know what usually takes longer, what usually takes shorter, and where risk occurs. Scope creep, integration issues, resource allocation - all common problems to run into. They’ll also know what to do when a problem occurs, and how to address them quickly.

A great first step is to just come to an estimate, even if it includes variability. A well-defined statement of work can solve for this, and real-time visibility using project management tools like Monday helps keep expectations aligned.

The inevitable next question is…what happens if delays happen? What happens if BIG delays happen?

This is another issue you can agree to tackle when it comes up. Regular communication should be par for the course, so discuss these in your Slack channel or during your weekly call. 

There are a few ways to concretely handle this, though, which we’ll talk about next.

Discuss what to build and what to buy

If you’re reading this, you’re probably not Facebook or OpenAI, who devote billions of dollars to train their own AI models from scratch.

So, at some point, you’ll be working with someone else’s solution - be it cloud compute, AI models, or data lakes.

But there’s a layer right below that where you might run into issues trying to build your own solution. It could be at the start, realizing that adding functionality in a key component balloons cost or development time. 

Or it could be during the project - you’ve been working on that component for a few weeks, and are realizing it will take more $$$ to get it working than you’d like to commit.

Either way, the discussion of outsourcing technical solutions inevitably comes up. There needs to be a conversation - discussing tradeoffs of continuing to develop your own solution vs outsourcing.

Answer questions around:

  1. Remaining timeline and budget
  2. Feature set of existing off-the-shelf solutions
  3. Long-term opportunity cost of relying on an external vendor
  4. Priorities for the project (time, money, features)

This can help you make decisions about getting unstuck later on.

Set expectations at the start

Successful AI projects are narrow in scope. Instead of asking an AI solution to handle everything, define specifically what you want it to do. Going from broad panacea to “offload this step in the process to AI” is how the discussion should start. 

“The expectation is that the LLM is just like a person, and can do everything super quickly. But expecting that much out of a system that gets expensive - if you focus on what’s necessary, then the system can still perform really well. The idea is to simplify.” says Jurgis Samaitis, an MLOps Engineer at NineTwoThree. 

This is especially true if the idea started as a flashy hackathon project or demo. Those are great as a proof-of-concept, but a working demo != a working project, and setting expectations at the start helps everyone understand what success looks like.

Define what process or job the AI will perform. Especially if this solution is customer facing - if any AI-generated output will reach the customer, there need to be guardrails established to make sure it represents the brand well.

Answer questions like:

  1. Where will the AI solution exist in the process?
  2. How critical are cost, latency, and accuracy? In what order are they important?
  3. Will this solution be customer-facing? In what form? Is it mission-critical for the AI and its output to behave in a specific way?
  4. How often should the AI re-train on new information?
  5. How will we measure the solution’s performance? What criteria will we use? Who will be testing it? What does safety look like?

Better yet, write down the solution scope at the beginning. This initial step is crucial. In fact, knowing WHEN to use an LLM is half the battle. 

Taking a smaller, modular approach - where the LLM is inserted at a specific point, for a specific purpose, with safety if something fails - is a great way to start.

Data quality is everything

It’s almost impossible to overstate how important data quality is to any successful AI project.

“You expect that the data is going to be clean, informational, and quality, but in nine out of ten cases, that’s not happening.” says Vitalijus.

The old saying goes “garbage in = garbage out”, and that’s no different here. The best-designed system, with incredible feedback loops and robust safeguards, will still perform terribly if the underlying data has issues.

There’s a few things to consider when it comes to quality data:

  1. What kind of data is it? What’s the background? What features are important?
  2. Where is the data stored? Is there a database we can access? If not, where does it sit?
  3. Is it accurate? What confidence level do we have that it’s accurate?
  4. Is there a proper data pipeline set up?
  5. How often will it be used? Do we need real-time?
“You need proper data pipelines as well,” says Vitalijus, “and you need to figure out extract-transform-load (ETL) processes first, before you even start training.”

It’s important to know what a good data pipeline is. To define that, let’s go over what “bad” looks like:

  1. Manual data upload processes - As a matter of fact, avoid anything manual. This should be as automated as possible.
  2. Zero quality checks - It’s harder to code your way out of empty fields, irrelevant columns, and null values, compared to cleaning the data at the start.
  3. Segmented, siloed data - Unifying data takes significant time, complexity, and cost. Investing in a data lake is the right way to solve your data problem.
  4. Zero scalability - Sure, the data pipeline might work for a POC, but scaling to production workloads shouldn’t cause any investigation bottlenecks.
  5. Minimal monitoring - If something breaks, you can’t find out once the customers are complaining about garbage AI responses. 

If you spend most of your time on this step, that’s not necessarily a bad thing.

A word of warning: this can feel like a tarpit for any machine learning project. It can take longer than you’d expect to figure out your data problems (and that’s why billion-dollar companies exist to solve that specific problem.) 

However, if you feel like you’re spending too much time on this step, you probably aren’t - that’s how critical it is.

Choosing models isn’t as critical…yet

Sort of. Most people think that which model you pick determines the outcome of the entire project. So they want whatever’s the best, and usually the most expensive. Think the latest Claude or GPT. This might be the best answer, but it comes at the cost of speed and price.

So sometimes, that’s helpful. But there are plenty of other models out there. 

We’ve found that picking the perfect model for the first pass through fine-tuning isn’t necessary. Great, cheap models like GPT-3.5 can teach us a lot about the performance and feature selection, without breaking the bank.

“Data is king, and feature engineering is king. If you’ve prepared your data in the right way, you can get to 80-90% accuracy using a cheap base model. That’s more than enough for a proof-of-concept.” says Vitalijus.

Yes, down the road, it’s better to dial in a model. Maybe a model trained specifically on code, if you’re working on AI that’s generating code.

But for testing at the start, speed and quick iteration is the most important consideration.

Clearly define what testing looks like

This is usually where a lot of friction and delay can happen. Most people skip over this in the planning phase, and expectations are misaligned on what “correct” looks like.

Think about it - most software projects have predictable outputs and specific outcomes. 

“One of the key points in terms of timely delivery is to remember that this is statistics. It’s predicting the next word, and it’s really good at that. But statistics is unpredictable in nature, and there are gray areas.” says Jurgis.

Statistics can be unpredictable - it’s a really good approximation, not a guarantee of an answer. So when you create a chatbot to answer questions, and the answer isn’t exactly the same each time, it can be surprising. It’s important to understand that fixing one response from an LLM might make ten other responses unreliable.

This is where it’s helpful to bring in the experts on your team. If you’re replacing a manual task, talk to the employees who won’t have to do the task anymore. They’re best-equipped to understand what “correct” and “adequate” look like in terms of performance

Agreeing on an evaluation suite can clear up confusion and communication cycles, too. 

  1. What happens when a change is made to the system? 
  2. How do we test that change to make sure it doesn’t break other components? 
  3. How do we evaluate safety against prompt injection?
  4. Can we reliably, consistently measure the accuracy of the answers? So reliably that we can know if new changes improved or degraded the answer quality?
  5. Can we create a repeatable set of test cases around these questions?
“When you have a lot of subsystems, and something hallucinates, you’re not going to be aware something went wrong until you have very bad feedback from the customer. So it’s important to have robust evaluations at the start, to compare the output of every iteration.” says Vitalijus.

If you didn’t do this earlier, do it as soon as possible now.

…and what “done” is

We speak from experience - there’s a point where adjustments hit diminishing returns on these projects.

R&D improvements in machine learning are not linear. Getting from 0% to 95% complete can often be straightforward, but it can take just as long to get from 95% to 98% (and forget about 100%.)

“Achieving 100% accuracy is impossible for most ML use cases. We live in a non-deterministic world, and while data reveals patterns, there are almost always outliers and unexpected events that impact predictions.” says Vitalijus.

Some PhD-level professionals spend years improving a single metric by a few percentage points - it’s that complex and difficult to do for non-deterministic ML products.

So, it’s important to define what “done” looks like, and revisit that definition often. If you’re building an internal tool to help your team, it’s likely already offloading a significant chunk of their work. That means it probably shouldn’t be perfect 100% of the time. 

That’s a dramatic difference from a public-facing tool that customers will use - they’re expecting an extension of the company, and performance and accuracy should reflect that.

Work with your team and the agency to determine what’s realistic for a complete project. And we truly mean realistic - 100% accuracy is extremely difficult to achieve for AI projects. 

That also means defining what ISN’T good enough. If the entire system hinges on doing one thing really well, then it should perform that one function extremely well. Standards should remain high, and that can reflect in the testing and integration phase.

Reflect on any legal concerns

It’d be irresponsible not to address this. There are two main parts to liability.

The first is ensuring your data is secure, with strict governance and access controls. This should be table stakes for any modern enterprise, but it’s especially crucial for sensitive data, like healthcare information or legal documents. 

Second is guardrails for data being exposed to the outside world. This is something to agree on early in the project. Is it a liability? If so, how much should we invest in ensuring it doesn’t happen? That affects your testing and design phase considerably.

Finally, consider cost at each stage

Spending more time, exploring solutions, upgrading models, testing…all adds considerable cost to any project. And if you have infinite money to throw at this, then this section isn’t for you.

But if you’re focused on building a sustainable business built on AI features, cost should be your number one concern. It’s easy to see costs balloon when chasing the last percentage of accuracy improvements, or upgrading models from GPT-4 to GPT-5.

This section could be its own article, but the main gist is simple: following the advice in the previous sections helps control cost. Increasing predictability, limiting scope, having upfront conversations to align everyone…that creates a nimble, streamlined project. 

And if it does come time to adjust, make it a team decision.

But sometimes, it’s helpful to see a real-world example.

Case study

NineTwoThree was selected by the CR Innovation Lab to help build an experimental chatbot that combines the power of AI with CR's expertise to answer your questions and offer product recommendations.

Read more here.

As you can see…

There’s a lot to consider if you want an efficient project. 

Partnering with experts is an excellent way to get a head start on your organization’s AI journey. Don’t let unrealistic expectations get in your way - communicate often, align on every step, and you’ll be well on your way.

In case it isn’t clear, we’ve got a lot of experience with this. If you’re interested in how AI can save your organization real money and time, let’s chat.

If you like this, download the full resource here.
PDF This Page
Key Elements for Timely Completion of Machine Learning Projects
View this Resource as a FlipBook For Free
Key Elements for Timely Completion of Machine Learning Projects
Download Now For Free
contact us

Have a Project?
Talk to the
Founders Directly

It's free, what do you have to lose?