Hiring Guide

Questions to Ask an AI Agency Before You Hire

Vetting an AI agency? These 35 questions separate competent partners from vendors who overpromise and underdeliver.

Published March 06, 2026

Most companies ask the wrong questions when evaluating AI agencies. They focus on "What AI technologies do you use?" (who cares, they all use the same tools) instead of "How do you handle model performance degradation in production?" (this is where projects fail). This guide gives you the questions that actually reveal whether an agency can deliver.

I've organized these questions into categories. You don't need to ask every single one—pick the ones most relevant to your project type and risk tolerance. But every category matters. An agency that gives great answers on technical capability but vague responses on project management will leave you frustrated even if they deliver good technology.

Technical Capability Questions

1. What's your process for selecting which machine learning approach to use for a given problem?

You want to hear: systematic experimentation, benchmarking multiple algorithms, comparing results against baseline metrics. Red flag answers: "we always use neural networks" or "we pick based on what's trendy."

2. Walk me through a project where your first technical approach didn't work. What did you do?

Good agencies experiment, hit dead ends, and pivot. If they claim 100% success on the first try for every project, they're lying. Look for honest stories about failed experiments, lessons learned, and how they adapted.

3. How do you handle imbalanced datasets?

This is a common real-world problem (fraud detection, defect identification, rare disease diagnosis). Strong answers mention oversampling, undersampling, synthetic data generation (SMOTE), cost-sensitive learning, or adjusting decision thresholds. If they look confused, they haven't done much real work.

4. What's your approach to model interpretability and explainability?

Critical for regulated industries or high-stakes decisions. Good agencies talk about SHAP values, LIME, attention mechanisms, feature importance, or simpler interpretable models. If they say "neural networks are black boxes, just trust them," walk away.

5. How do you prevent overfitting?

This is a basic ML question. Answers should mention train/validation/test splits, cross-validation, regularization techniques, early stopping, or ensemble methods. Any agency that can't answer this clearly lacks fundamental expertise.

6. Describe your deployment architecture for a typical ML system.

You're looking for: API layers, monitoring infrastructure, A/B testing capability, rollback procedures, version control for models, separation between training and inference, security considerations. Vague answers like "we deploy to the cloud" are insufficient.

7. How do you handle model retraining and versioning in production?

Models degrade over time. Good agencies have systems for scheduled retraining, detecting data drift, tracking model versions, and gracefully switching between model versions without downtime.

8. What tools and frameworks do you typically use, and why?

The specific tools matter less than the reasoning. An agency that says "we use TensorFlow because it's popular" is less impressive than one that says "we use PyTorch for research projects because of its debugging capabilities and TensorFlow for production because of its deployment ecosystem."

Process and Project Management Questions

9. What does your discovery or scoping phase include?

Strong agencies do paid discovery before committing to a full project. This should include stakeholder interviews, data audits, feasibility analysis, architecture design, and a detailed project plan with milestones.

10. How do you prioritize which AI use cases to tackle first?

Look for frameworks that balance business value, technical feasibility, data readiness, and strategic alignment. Agencies that just build whatever you ask for without pushing back are order-takers, not partners.

11. What's your typical project timeline from kickoff to production deployment?

For context: simple automation projects take 2-3 months, standard ML projects take 4-6 months, complex custom systems take 6-12 months. If an agency promises a sophisticated AI system in 4 weeks, they're either reusing an existing solution or setting unrealistic expectations.

12. How do you handle scope changes during a project?

Scope changes are inevitable. Good agencies have a formal change request process, transparent pricing for additional work, and clear communication about how changes affect timeline and budget.

13. What does your typical project team structure look like?

You want to hear about multiple roles: ML engineers, data engineers, project managers, potentially domain experts. A single person wearing all hats is a red flag for any non-trivial project.

14. How often will we have check-ins, and what format do they take?

Weekly syncs are standard for active projects. Monthly check-ins suggest you'll be in the dark most of the time. Ask about demo frequency, status reporting, and escalation procedures.

15. What happens if you miss a milestone or deliverable?

Honest agencies acknowledge this possibility and have contractual provisions for delays, remediation, or partial refunds. Agencies that claim they never miss deadlines are either lying or have only done tiny projects.

Data and Security Questions

16. How do you handle data security and privacy?

Crucial if you're sharing customer data, health records, financial information, or trade secrets. Good answers mention encryption (in transit and at rest), access controls, compliance certifications (SOC 2, ISO 27001), data minimization, and anonymization techniques.

17. Will our data be used to train models for other clients?

The answer should be "no" unless you explicitly agree otherwise. Agencies can learn general techniques from your project, but your actual data should never train models for competitors.

18. Where will our data be stored and processed?

Important for compliance (GDPR, HIPAA, industry-specific regulations). You need to know which cloud regions, whether data ever leaves your infrastructure, and who has access.

19. What's your approach to data quality assessment?

Data quality makes or breaks AI projects. Strong agencies audit data completeness, accuracy, consistency, and timeliness before building anything. They should show you data quality reports early in the project.

20. How do you handle missing, incorrect, or inconsistent data?

This is reality for most business datasets. Look for answers about imputation techniques, validation rules, outlier detection, and working with stakeholders to improve data collection processes.

Business and Commercial Questions

21. Can you share references from similar projects?

Always ask. Talk to at least two references. Ask them about communication, deadline adherence, quality of deliverables, and how the agency handled problems.

22. What IP and code ownership terms do you offer?

You should own all code, models, and documentation produced for your project. Some agencies try to retain ownership of "core IP" or frameworks—negotiate this upfront.

23. What's included in your pricing, and what costs extra?

Make this explicit. Does the quote include data pipeline development? Cloud hosting? Post-launch support? Model retraining? Surprises here kill budgets.

24. What does post-launch support look like?

Agencies should offer some warranty period (30-90 days) for bug fixes, plus optional maintenance contracts for ongoing support, monitoring, and updates. Get pricing for both upfront.

25. How do you bill for scope increases or additional work?

Fixed-price projects need clear change order processes. Time-and-materials projects need transparent timekeeping. Understand the mechanism before signing.

26. What happens if we want to end the engagement early?

Shit happens. Companies get acquired, priorities change, budgets get cut. You need a contract that lets you exit gracefully, ideally with a 30-day notice and payment only for work completed.

27. Will we own the trained models and can we move them to a different provider?

Model portability matters if you outgrow the agency or bring work in-house. Models should be delivered in standard formats (ONNX, SavedModel, pickle) that aren't locked to proprietary platforms.

Industry and Domain Expertise Questions

28. Have you built systems for companies in our industry?

Industry expertise accelerates projects. An agency that's built fraud detection for three fintech companies understands your data, regulations, and common pitfalls. An agency doing their first fintech project will learn on your dime.

29. What industry-specific regulations or compliance requirements have you handled?

Healthcare (HIPAA), finance (PCI-DSS, FINRA), government (FedRAMP), or EU operations (GDPR) all impose constraints on AI systems. Agencies with relevant experience save you compliance headaches.

30. Can you show us case studies or demos from similar use cases?

Case studies reveal how the agency thinks about problems like yours. Look for quantified outcomes, technical details, and honest discussion of challenges—not just marketing fluff.

Red Flags to Watch For

31. What's a project that didn't go well for you, and what did you learn?

Agencies that claim perfect track records are either new, lying, or unwilling to be honest with you. The best agencies own their failures and explain how they've improved.

32. How do you stay current with AI developments?

AI evolves fast. Strong agencies invest in R&D, attend conferences, publish research, contribute to open source, or run internal learning programs. Agencies that don't have a systematic approach to learning fall behind.

33. What percentage of your projects actually make it to production?

Some agencies specialize in proofs-of-concept that never ship. Others build production systems. Neither is wrong, but you need to know which one you're hiring. If you want a production system, hire an agency where 70%+ of projects deploy.

34. Have you ever recommended a client not pursue an AI project? Why?

Great agencies turn down bad-fit projects. If an agency has never said "no" to a potential client, they either have low standards or desperately need revenue.

35. What makes your agency different from competitors?

This is a softball, but the answer reveals what they value. Do they compete on price (commodity shop), specialization (deep expertise in a niche), or methodology (unique process or technology)? Understand their positioning.

How to Use These Questions

Don't turn the initial call into an interrogation. Pick 8-10 questions most relevant to your situation and weave them naturally into conversation. Listen for specificity, honesty, and depth of thinking. Great agencies give concrete examples, acknowledge trade-offs, and ask you smart questions in return.

Follow up with harder questions in later conversations as you narrow your shortlist. References, contract terms, and commercial questions come after you've established basic technical fit.

Evaluating the Answers

You're not just listening to what agencies say—you're evaluating how they say it:

Depth vs. handwaving: Do they explain their reasoning or throw around buzzwords? Can they go deeper when you probe? Do they admit when something is hard?

Confidence vs. arrogance: Do they acknowledge uncertainty and trade-offs, or claim to have perfect solutions for everything?

Client focus: Do they ask questions about your business, constraints, and goals? Or do they just pitch their standard offering?

Transparency: Are they upfront about costs, timelines, and risks? Or vague and evasive when you ask for specifics?

The best agencies earn your trust by demonstrating competence, honesty, and curiosity. They treat you as a partner, not a transaction.

After You Ask the Questions

Once you've vetted 3-5 agencies, ask for proposals that include:

  • Detailed technical approach
  • Project timeline with milestones
  • Team composition and roles
  • Pricing breakdown
  • IP ownership terms
  • Post-launch support options

Compare proposals on equivalent scope, not just price. The cheapest option usually excludes things the more expensive proposals include (data pipeline work, deployment, support, documentation). Normalize the proposals so you're comparing apples to apples.

The Meta-Question

Here's the most important question, even though you don't ask it directly: Would I want to work with this team for 6 months?

AI projects are collaborative. You'll have weekly calls, make decisions together, and solve problems as a team. Technical capability matters, but so does communication style, responsiveness, and cultural fit. Trust your gut—if something feels off during the sales process, it won't get better during execution.

Ready to start vetting agencies? Browse our directory of 700+ verified AI agencies, read their case studies, and compare their expertise to find the right fit for your project.

Ready to Find the Right AI Agency?

Browse 700+ verified AI agencies. Filter by tech stack, industry, location, and client ratings.

Browse AI Agencies