Automation

Can AI Convert Natural Language Into Structured Business Data?

Marius Jeskulke
Marius Jeskulke · Partner
·8 min read

A sales rep comes out of a client meeting, picks up their phone, and speaks: "Quote for Meier GmbH, 50 climate units model ProAir 3000, delivery by end of June, discount same as last time." A few minutes later, a structured quote sits in the system. The article number was looked up in the product catalogue, the discount pulled from the customer history, the delivery date checked against current stock.

This is not a future vision. It works today because over the past two years, the way business software processes input has fundamentally changed. Instead of filling out forms, instead of navigating the right screen in the right sequence, you speak or type in natural language. AI translates that into structured data and actions.


The real capability is not speech recognition

When people think about voice input in business software, they think of speech recognition first. Microphones, transcription, dialects. That's the wrong starting point. Speech recognition is a solved problem. Current models recognise regional accents as reliably as standard pronunciation, and even domain-specific terminology can be corrected through a post-processing step that reconciles the transcript with a specialised glossary.

The real capability is something else: understanding natural language and translating it into structured actions. "Quote for Mueller" becomes a customer lookup. "200 units Classic" becomes an article assignment. "Discount same as last time" becomes a customer history query. This isn't a dictation device -- it's an agent that understands what's meant and acts on it.

From spoken command to structured action: Natural input becomes customer lookup, article assignment, and history query. An agent that understands and acts, not a dictation device.
From spoken command to structured action: Natural input becomes customer lookup, article assignment, and history query. An agent that understands and acts, not a dictation device.

Voice and text are interchangeable input channels

An insight that has solidified across our projects over months: voice and text are simply different input channels for the same agent. The language model is, at its core, a user interface. Behind it run the same backend agents, the same business logic, the same system integrations. Whether someone types a request or speaks it makes no difference to the system.

This fundamentally changes project planning. You don't build a voicebot first and then a chatbot. You build the business logic once and then put any input channel in front of it. An agent that processes written requests can, with manageable effort, also be operated by voice. The reverse works just as well.

In practice, this means: if a system works via chat, the step to voice input is not a new project -- it's an extension of the existing one. The investment in business logic pays off twice over.


Where voice input beats forms

Forms force the user into a predefined structure. That works as long as the input is unambiguous. But many business processes are not unambiguous.

"Create a quote for Mueller." Which Mueller? A form would need a dropdown with all customers named Mueller. Or a search field with filters. The user would need to know that they first have to search for the customer, then verify the address, then select the product category. An agent with access to customer data finds eight matches, infers from context that it's probably Mueller GmbH in Stuttgart, and only asks when it's genuinely unclear.

Voice input works especially well wherever a process has decision points that aren't all known in advance. A form has to anticipate every possible path. An agent can react flexibly, ask follow-up questions, and use context to make the right decision.

The flip side: for structured, predictable input, a well-designed form remains more efficient. The question is not "voice or form" but "where in the process does each input method pay off?"

Form vs. agent: Forms need every path in advance. An agent reacts flexibly, uses context, and only asks when there is genuine ambiguity. For "quote for Mueller," the agent finds the right one among eight matches.
Form vs. agent: Forms need every path in advance. An agent reacts flexibly, uses context, and only asks when there is genuine ambiguity. For "quote for Mueller," the agent finds the right one among eight matches.

The most common use case: automating standard requests

The most obvious application is repetitive standard requests. Price and delivery time enquiries by phone, standard support questions, recurring orders. At one industrial company, we found that the majority of incoming calls had the same structure: customer states article number, asks about price and availability.

Automating such requests works because they are predictable. The AI understands the request, asks for clarification when needed, reconciles the information against the ERP system, and delivers an answer. The employee is freed from repetitive work and can focus on the complex cases that genuinely require human judgement.

Experience shows: companies that automate standard requests don't just gain time. They also gain capacity for the cases that make the difference against the competition.


Knowledge becomes tangible because AI demands it

A side effect that turns out to be one of the most valuable: introducing natural language as an input channel forces organisations to articulate their implicit knowledge.

When a sales rep simply speaks their request and the system is supposed to make the right decision, the system has to know what the right decision looks like. What discount tiers exist? Under what conditions does which price apply? Which products are compatible? In many companies, all of this is knowledge that lives in the heads of individual people but is documented nowhere.

AI forces this documentation because it cannot make good decisions without explicit rules. This isn't a documentation project in the classic sense -- one that runs alongside daily operations and eventually becomes outdated. It's a knowledge base that is actively used and gets better with every correction.


What voice input is not: a universal default path

Voice input in an open-plan office doesn't work. Too loud, too little privacy, socially inappropriate. In a home office or in the car, on the other hand, it's the most natural input method. The decision for or against voice input is not a technology question -- it's a context question.

There is also a generational component: people who are used to communicating via voice messages find voice input in software natural. People who mostly type prefer the chatbot. A good system offers both and leaves the choice to the user.


When AI mishears: confirmation instead of blind trust

What happens when speech recognition gets an article number wrong? When "156755" suddenly becomes "156756" and the wrong product ends up in the quote?

The solution is the same one a good call centre employee would use: confirmation of critical information. "Did I get that right -- article number 156755?" On top of that, the AI checks semantically whether the recognised number matches the described product. If a customer asks for a climate unit but the recognised article number leads to a heating controller, there is clearly an error.

Confirmation for critical data: From voice input through transcription and plausibility check to confirmation. The same principle applies to written input.
Confirmation for critical data: From voice input through transcription and plausibility check to confirmation. The same principle applies to written input.

This is not a voice-specific problem. It's general agent design: important information is confirmed and plausibility-checked. The same principle applies to written input when a user makes a typo.


Data protection: pragmatic, not perfect

Voice processing in a business context has a legal dimension. Voice data is biometric data under GDPR and requires conscious handling.

In practice, a pragmatic approach has proven effective: voice data is deleted immediately after transcription. No audio is stored, only the transcribed text. This significantly reduces the data protection attack surface, because biometric characteristics can no longer be derived from a transcript.

For meeting transcriptions, an additional rule applies: always ask beforehand. Experience shows that acceptance is high when you communicate transparently that the meeting is being transcribed, not recorded. "We use the transcription so we can process the meeting outcomes more effectively" is an argument that works in practice with clients of all sizes.


The architecture: three layers, cleanly separated

Building a natural language interface for business software requires three layers:

Input layer: Voice or text is converted into a uniform text format. For voice, this means: transcription, post-processing with a domain glossary, optionally confirmation of critical data. For text, this step is skipped.

Understanding layer: The agent interprets the input, determines the intent, and identifies the relevant parameters. "Quote for Mueller, 200 units Classic" becomes: action = create quote, customer = Mueller GmbH, product = Classic, quantity = 200. In cases of ambiguity, the agent asks.

Execution layer: The agent carries out the identified action. Customer lookup in CRM, article query in ERP, price calculation, quote generation. This layer is identical regardless of whether the input came via voice or text.

The clean separation of these layers is the key to maintainability. If speech recognition improves, you swap out the input layer. If a business process changes, you adjust the execution layer. The layers don't affect each other.

Three-layer architecture: Input layer (voice or text), understanding layer (recognise intent and parameters), execution layer (actions in ERP, CRM, etc.). Cleanly separated, independently replaceable.
Three-layer architecture: Input layer (voice or text), understanding layer (recognise intent and parameters), execution layer (actions in ERP, CRM, etc.). Cleanly separated, independently replaceable.

Not everything needs voice

Voice input is a tool, not a goal. Some processes benefit from it, others don't.

Well suited: Data capture after client meetings (faster than typing), standard enquiries by phone (price, delivery time, availability), documentation in the field (hands free), and situations where the user doesn't know which fields to fill in (guided dialogue instead of a rigid form).

Less suited: Highly structured input with many technical parameters (a form is faster), noisy environments, and processes where input must be exact and not open to interpretation (e.g. calculations).

The decision should be made per process step, not across an entire piece of software.

Where voice input pays off and where it doesn't: Well suited for data capture after meetings, standard enquiries, field work, and unfamiliar form structures. Less suited for many technical parameters, noisy environments, and exact input.
Where voice input pays off and where it doesn't: Well suited for data capture after meetings, standard enquiries, field work, and unfamiliar form structures. Less suited for many technical parameters, noisy environments, and exact input.
A system that enables initial capture by voice and detail editing by form combines the strengths of both approaches.
Marius Jeskulke
About the author
Marius Jeskulke
Partner

Marius Jeskulke brings 20 years of experience in software development — from developer through CTO to entrepreneur. Today, with Deyan7, he supports mid-sized companies in the value-driven integration of AI.