Monology Logo
Back to Blog
AI & Machine LearningFeatured

95% Intent Classification: How We Trained AI to Separate Clients from Job Seekers

Behind the scenes look at how we built an AI system that achieves 95% accuracy in distinguishing between client inquiries and job applications - solving the hidden cost of mixed contact forms in IT services.

AI Enthusiast

8 min read
#AI#Machine Learning#Intent Classification#BERT#Lead Management#Sales Automation#IT Services#Open Source
Featured image for 95% Intent Classification: How We Trained AI to Separate Clients from Job Seekers

The $50K Misunderstanding That Started It All

Last month, our sales team spent three days preparing a comprehensive proposal for what they thought was a major enterprise client requesting React development services. The inquiry came through our contact form: "We need experienced React developers for our upcoming project."

After hours of research, competitive analysis, and proposal preparation, we discovered the truth during the first call - they weren't looking to hire our development services. They were a recruiting agency looking to poach our React developers.

This wasn't an isolated incident. Our analysis revealed that 23% of inquiries through our main contact form were misclassified, leading to wasted sales effort, confused prospects, and missed opportunities. The solution? An AI-powered intent classification system that now achieves 95% accuracy in separating genuine client inquiries from job seekers, recruiters, and other non-client contacts.

The Hidden Cost of Mixed Inquiries

Most IT service companies face this challenge without realizing its true impact:

  • Sales Time Waste: Our analysis showed sales reps spent an average of 2.3 hours on misclassified leads before discovering the error
  • Response Delay: Genuine client inquiries got delayed responses while the team chased false leads
  • Opportunity Cost: During peak periods, misclassification caused us to miss follow-up windows on real prospects
  • Team Morale: Constant false leads frustrated our sales team and reduced their effectiveness on genuine opportunities

The business impact was clear: we needed an automated system to classify inquiries before they reached human hands.

Building the AI Classification System

We developed a BERT-based machine learning model specifically trained on IT services inquiries. Here's how we tackled the technical challenges:

The Training Data Challenge

Creating a high-quality dataset was our first hurdle. We needed examples that captured the subtle differences between client inquiries and job-seeking attempts. Our final training dataset included:

  • 70,000+ labeled examples across 7 intent categories
  • 10,000+ samples per intent to ensure balanced learning
  • Real-world variations including typos, informal language, and industry jargon
  • Edge cases that commonly caused human misclassification

The key categories we trained the model to recognize:

  • Requirement Submission (client project needs)
  • Job Application Submission (candidates seeking employment)
  • General Query (information requests)
  • Contact Details (wanting to reach specific people)
  • Feedback Submission (reviews and suggestions)
  • Appreciation (thank you messages)
  • Greeting (initial contact attempts)

The Technical Architecture

We chose BERT (Bidirectional Encoder Representations from Transformers) for its superior performance on text classification tasks. The model architecture includes:

  • BERT-base-uncased as the foundation model
  • Custom classification head trained on our specific intents
  • Context-aware processing that considers entire message content
  • Real-time inference with sub-50ms response times

Training took approximately 45 minutes on modern GPU infrastructure, with the model achieving 95.2% accuracy on our held-out test set.

The Tricky Cases: Common Misclassification Patterns

During development, we discovered several patterns that commonly fooled both humans and early model versions:

The Ambiguous "We Need" Pattern

These messages were the trickiest to classify:

  • Client Intent: "We need React developers for our e-commerce platform rebuild"
  • Recruiting Intent: "We need React developers for our client's fintech startup"

The difference? Clients talk about their projects, while recruiters mention client projects or use possessive language that distances them from the actual work.

The Service vs. Employment Confusion

Another common pattern:

  • Client Intent: "Looking for cloud migration services"
  • Job Seeker Intent: "Looking for cloud migration opportunities"

Our model learned to distinguish between "services" (what clients buy) and "opportunities" (what job seekers want).

The Formal vs. Informal Language Trap

Job seekers often use more formal language:

  • Job Seeker: "I would like to inquire about potential career opportunities in your organization"
  • Client: "Need help with our API integration ASAP"

Clients typically use more direct, business-focused language with urgency indicators.

Context Clues That Make the Difference

Our AI system learned to identify subtle linguistic patterns that indicate true intent:

Ownership Language

  • Client indicators: "our project," "our team," "our deadline"
  • Non-client indicators: "their requirements," "client needs," "candidate search"

Urgency and Timeline Markers

  • Client patterns: "ASAP," "urgent," "deadline," "go-live date"
  • Job seeker patterns: "when convenient," "at your earliest opportunity," "flexible timeline"

Technical Specificity

Genuine clients often include specific technical details about their existing systems, while job seekers focus on their skills and experience.

Real-World Performance Results

After six months in beta, our intent classification system has delivered measurable results:

  • 95.2% classification accuracy on live inquiries
  • 67% reduction in sales time wasted on non-client inquiries
  • 34% faster response time for genuine client inquiries
  • 23% increase in sales team satisfaction scores
  • 18% improvement in lead-to-opportunity conversion rate

Implementation Lessons Learned

Data Quality Beats Quantity

We initially thought we needed millions of examples. In reality, 10,000 high-quality, diverse examples per intent category proved sufficient for production-level performance.

Context Windows Matter

Single sentences often lack sufficient context. Our model performs best when analyzing the complete inquiry, including subject lines and signatures.

Continuous Learning Is Essential

We retrain the model monthly with new examples, especially edge cases that initially caused misclassification. This continuous improvement maintains accuracy as language patterns evolve.

Human-in-the-Loop Validation

For inquiries where the model confidence is below 85%, we route them to human review. This catches the remaining 5% of edge cases while maintaining efficiency.

The Open Source Solution

Recognizing that many IT service companies face similar challenges, we've open-sourced our intent classification system. The complete solution includes:

  • Pre-trained BERT model achieving 95%+ accuracy
  • 350+ sample training dataset
  • Google Colab training notebook for customization
  • Production-ready inference code
  • Integration examples and documentation

You can access the complete system at: https://github.com/monology-io/intent-classification-bert

The repository includes everything needed to train your own model with your specific data, or use our pre-trained version as a starting point.

Beyond Basic Classification

Our current system goes beyond simple client vs. non-client classification. It routes inquiries to specialized teams based on detected intent:

  • Technical requirements → Solutions architects
  • General inquiries → Customer success team
  • Urgent requests → Priority response queue
  • Job applications → HR department
  • Partnership inquiries → Business development

This intelligent routing ensures every inquiry reaches the right person immediately, improving response quality and customer experience.

The ROI of Smart Classification

Six months after implementation, the numbers speak for themselves:

  • Cost savings: $25K annually in reduced sales time waste
  • Revenue impact: $45K in additional deals closed due to faster response times
  • Operational efficiency: 40% reduction in inquiry processing time
  • Customer satisfaction: 28% improvement in response relevance scores

The system paid for itself within 60 days of deployment.

What's Next: Advanced Intent Understanding

We're already working on the next version with enhanced capabilities:

  • Sentiment analysis to identify urgent or frustrated inquiries
  • Company size detection to route enterprise vs. SMB inquiries appropriately
  • Technology stack identification to match inquiries with specialized team members
  • Budget range estimation based on project scope indicators

Getting Started Today

If you're losing deals to misclassified inquiries, you don't need to build from scratch. Our open-source solution can be deployed in under an hour:

  1. Download the pre-trained model from our GitHub repository
  2. Test with your existing inquiries to validate performance
  3. Customize with your data using our Google Colab notebook
  4. Integrate with your contact forms using our API examples
  5. Monitor and improve with our continuous learning approach

The era of manually sorting through mixed contact form submissions is over. AI-powered intent classification isn't just a nice-to-have anymore—it's a competitive necessity for IT service companies serious about efficient lead management.

Your sales team's time is too valuable to waste on misclassified inquiries. Let AI handle the sorting so your humans can focus on closing deals.

Navanath Jadhav profile picture

Navanath Jadhav

AI Enthusiast

Passionate about practical AI applications in business automation. Specializes in natural language processing and intent classification systems. Believes in open-source solutions and sharing knowledge to help companies leverage AI effectively. When not training models, enjoys exploring the latest developments in transformer architectures and their real-world implementations.