Sequoia Applied Technologies logo Case Study

Last updated: March 4, 2026

Teams want faster cycle time and better coverage, yet privacy and compliance matter. SequoiaAT uses careful AI setups to raise speed while keeping control of code and data.

Client stories

Faster test case creation with AI and LLMs

Challenge: manual work slowed releases and raised costs. Concern about code privacy.

  • Use multiple models for idea generation on test cases for safer parts of the app.
  • Tune prompts and guardrails for product context and naming.
  • Keep review in the loop for final acceptance.

AI created data for automation at scale

Challenge: static datasets limited coverage and missed edge cases.

  • Plug a large language model into the test farm to supply fresh inputs on demand.
  • Include rare and boundary inputs to stress the system.
  • Track pass and fail trends to guide next runs.

Triage insights that improve coverage

Challenge: field issues surfaced patterns that test sets did not cover well.

  • Use ML to compare triage notes with the test library and find gaps.
  • Suggest new cases that mirror real issues seen in the wild.
  • Fold results into the plan for the next sprint.

Pilot of offline AI for unit tests

Challenge: speed up unit test authoring without sending code outside the network.

  • Run an offline model like Llama inside the build setup.
  • Limit scope to selected repos and keep developer review in place.
  • Measure speed, flakiness, and real fault catch rate before scale up.
Book a working session

How Sequoia approaches AI in testing

Most teams that try AI for testing run into the same early friction. The tooling is not the hard part. The harder question is where to point it, how to protect sensitive code, and how to keep a human in the loop so the output is actually trustworthy.

Sequoia starts every engagement by mapping what can and cannot leave the network. Code that handles payments, health data, or authentication typically stays on-premise. Less sensitive modules, utility layers, configuration logic, and UI flows are often tractable candidates for external model assistance. That boundary is agreed before any tooling is configured.

Once the scope is clear, prompts are tuned to the product domain. A generic LLM asked to write tests for a medical device workflow will produce generic output. When the prompt is shaped around actual component names, known edge cases, and the team's naming conventions, the results become something a reviewer can work with quickly rather than rewrite from scratch.

Human review stays mandatory at every stage. AI-generated test cases go through the same approval cadence as anything else entering the suite. The goal is not to remove the engineer from the loop but to cut the time they spend on the parts of the job that are mostly mechanical.

Closing coverage gaps with real-world data

One of the more useful applications Sequoia has found is using production triage data to find gaps in existing test libraries. When the same class of defect appears in the field repeatedly, it usually means the test suite did not anticipate that pattern.

Comparing triage notes against the test library with an ML layer surfaces those blind spots in a way that manual review rarely does at scale. The output is a list of suggested test scenarios, mapped to the part of the product where the gap exists, which the team can review and prioritise in the next sprint.

Static datasets are the other common bottleneck. Automation suites that run against the same fixed input files miss boundary conditions, rare formats, and malformed inputs that only show up in real usage. Feeding an LLM into the test data generation step means the suite gets a wider variety of inputs on every run without the team having to curate them by hand.

This is not about replacing a proper test strategy. It is about giving the existing strategy more surface area with less manual effort.

Code privacy and data handling

This is the question that comes up in most conversations before anything else. Teams are right to ask it.

For clients where code confidentiality is non-negotiable, Sequoia runs models offline inside the build environment. Llama and similar open-weight models can be deployed on-premise with no external API calls. The model never sees anything outside the internal network.

For clients where some external model use is acceptable, the scope is limited to the parts of the codebase that have been explicitly cleared. Configuration, instrumentation, and test scaffolding typically qualify. Core business logic and proprietary algorithms typically do not.

Test data generated by AI is synthetic by design. It is never derived from real customer records. Where the product handles regulated data, the synthetic generation process is also reviewed against those requirements before use.

Common questions

Does using AI for testing mean less developer involvement?

No. Developer review remains part of every step. AI output, whether test cases or test data, goes through the same approval process as anything else entering the suite. The aim is to reduce time on mechanical work, not to remove judgment from the process.

Can this work if our code cannot leave the network?

Yes. For teams where code confidentiality is a hard requirement, Sequoia configures offline models that run entirely inside the build environment. No external API calls are made and the model operates only within the agreed scope.

How is scope defined at the start of a project?

The first step is a scoping session where Sequoia maps the codebase against sensitivity. Parts that handle regulated data, payments, or proprietary logic are typically kept outside AI-assisted workflows. The boundary is agreed in writing before configuration begins.

What does a pilot typically involve?

A pilot usually covers one clearly bounded part of the product over a short sprint. The team measures test creation time, suite stability, and defect catch rate against the baseline before deciding whether and how to expand scope.

Related services