7 or so AI Killer Prompts to Change Your Daily Life as a Functional Consultant

In this 112th Areopa webinar, Jose Miguel Azevedo shares a set of AI prompts he has built up over a year of daily use as a Business Central functional consultant at KPMG London. Moderator Bert Verbeek hosts the session. Jose covers six concrete prompts — each tied to a real consulting task — and closes with a set of bonus ideas and practical takeaways about how to use GenAI responsibly in consulting work.

Title slide: 7 or so AI killer prompts to change your daily life as Functional Consultant — Jose Miguel Azevedo, Areopa Webinar 112, September 29 2025
▶ Watch this segment

Why Functional Consultants Should Look at AI Now

Jose opens with a personal framing exercise: listing the tasks he enjoys versus the ones he tries to avoid. Repetitive documentation work, starting from a blank page, and lack of inspiration feature on the “avoid” side. He used that list as his starting point for finding where GenAI could help — focusing first on the things he disliked, rather than the things he was already good at.

He maps the main functional consultant responsibilities — requirements gathering, gap analysis, solution design, configuration, data migration, integration, testing, training, and documentation — and shows where AI can add value in each area.

Slide listing the main tasks of a Functional Consultant: Requirement Gathering, Gap Analysis, Solution Design, System Configuration, Data Migration, System Integration, Testing, Training, Documentation
▶ Watch this segment

Prompt 1 — Requirements Gathering from Meeting Transcripts

The pre-work matters as much as the prompt itself. Jose’s recommended steps: record the requirements session (Teams does this automatically), copy the transcript to a text file, do a quick manual review to remove contradictions and off-topic sections, and decide on the output format before you write the prompt.

The key difference between a weak and a strong prompt here is context. A bare “generate requirements from this transcript” gives AI nothing to prioritise. Jose’s version includes his role, the customer industry and geography, who attended the meeting and their titles, and whether the output should be a bullet list, Word document, or DevOps backlog item.

Prompt 1 slide showing the GenAI prompt structure for requirements gathering: role context, customer background, meeting details, attendee titles, and the final ask to generate requirements
▶ Watch this segment

Version 2 of the prompt asks for output formatted directly for Azure DevOps import, specifying columns: Work Item Type (always “Requirement”), Title, Description, State (always “Proposed”), Area Path (always “Finance”), and Tags (always “Business Analysis”). The result is a CSV that can be imported straight into a DevOps backlog without transformation.

Prompt 1 version 2 slide showing the DevOps-formatted output: columns Work Item Type, Title, Description, State, Area Path, Tags — with an APPROVED stamp and sample CSV rows
▶ Watch this segment
📖 Docs: Import work items from CSV — Azure Boards — step-by-step guide to the required columns and how to bulk-import requirements directly into your DevOps backlog.

Prompt 2 — Gap Analysis: Standard or Customisation?

Jose’s gap analysis prompt is deliberately minimal: describe the requirement precisely, and tell the AI what kind of answer you want — yes/no, or a list of options. He shares three real customer examples. Two of them (partial fixed asset disposal, and split invoice payment dates) get accurate, useful answers. The third — depositing a sales invoice to a bank account other than the company default — gets an incorrect answer; AI suggests customisation is required, when a recent Business Central update added that capability natively.

The lesson: use AI as a starting point for gap analysis, not as the final word. Always verify, especially for functionality added in recent release waves.

Prompt 2 slide showing an AI response to a gap analysis question about creating a sales invoice with partial payment dates, listing five approaches: Payment Terms Setup, Installment Payments, Deferral Schedules, Customization or Extensions, Manual Tracking
▶ Watch this segment
📖 Docs: About the Business Process Catalog — Dynamics 365 guidance — Microsoft’s catalog of 700+ standard business processes across Dynamics 365 apps, useful for benchmarking customer requirements against leading practice before or alongside an AI gap analysis.

Prompt 3 — Generating Functional Design Documents (FDDs)

FDDs need to serve multiple audiences at once: the customer for sign-off, developers for build guidance, and testers for test case derivation. Jose’s approach is to provide an agnostic FDD template (all section headings, no content), a clear description of the gap, and a set of high-level requirements as bullet points. The prompt asks AI to populate each section of the template for that specific gap.

In his example — a multi-company extension linking retailers to Territory Manager and District Manager tables — the AI handles abbreviations, scope, audience, assumptions, dependencies, pre-conditions, risks, and the functional design sections well. The one section it handles less well is Components; Jose notes this is an area to review manually.

Prompt 3 slide showing the full FDD prompt: role context, methodology, template sections (terminology, scope, audience, components, assumptions, dependencies, risks, functional design, technical notes, go-live, additional information), and high-level requirements
▶ Watch this segment

Prompt 4 — Test Case Generation from an FDD

Once you have a completed FDD, generating test cases is a single follow-up prompt: “Please propose the test cases for the previous FDD in the following format to be imported into Azure DevOps.” The format specifies Test Case Number, Test Case Title, and for each test case: Step Number, Action (text), and Expected Result (text).

Jose observes that few functional consultants enjoy writing test cases. The AI output covers both end-user steps and system validation steps at a level of detail that is immediately usable, reducing one of the most time-consuming documentation tasks in a BC project.

Prompt 5 — Anonymised Test Data for Data Migration

Before a data migration dry run you need test data that matches the configuration package schema without using real customer data. Jose’s prompt asks AI to generate 200 anonymised vendor records with specific columns matching a Business Central configuration package (No., Name, Search Name, Address, City, Phone No., Our Account No., Currency Code).

AI initially declines to produce 200 records in the chat window. After Jose pushes back, it offers a Python script using the Faker library that generates the file programmatically. The script works. Jose’s takeaway: the first answer is not always the best answer — challenge AI when the result falls short.

Prompt 5 slide showing the AI-generated Python script using the Faker library to produce 200 anonymized vendor records in CSV format matching the Business Central configuration package schema
▶ Watch this segment

Prompt 6 — Debugging Production Code Without Technical Skills

Jose describes a real post-go-live incident at a stock-market-listed group of companies. A low-code monitoring solution using Application Insights, BC Telemetry, Power Automate, and Job Queue was sending false-positive failure emails: the job queue would fail on its first attempt (often a deadlock), succeed on retry, but the alert still fired.

With his technical architect unavailable, Jose identified that the KQL query needed an additional condition — only send an alert when the number of failures equals the total number of allowed attempts — but could not compile the corrected query himself. On the third AI prompt attempt, the AI not only suggested the correct condition but also identified that the variable needed to be cast to integer using toint(), which was the root cause of the compile error.

Prompt 6 slide showing the corrected KQL query for Application Insights with the AI-suggested fix highlighted in orange: toint(customDimensions.alJobQueueNumberOfAttemptsToRun) == toint(customDimensions.alJobQueueExecutionNumberOfAttemptsToRun)
▶ Watch this segment

The solution has been running without issues for three years since. For a functional consultant with no AL or KQL background, AI provided an unblocking option that would otherwise have required waiting for a technical resource.

📖 Docs: Job queue lifecycle telemetry — Business Central — reference for the custom dimensions and KQL fields available when monitoring Job Queue entries via Application Insights.

Bonus Prompts

Jose closes the prompt section with additional ideas he uses but did not have time to demo:

  • Reviewing and rewriting important formal emails before sending
  • Drafting integration strategy and data migration strategy documents from a baseline template
  • Configuring Data Exchange Definitions in Business Central — one of the most tedious standard configuration tasks
  • Building a library of test scripts for ISV add-ons, where Microsoft does not ship test cases
  • Generating risks, assumptions, and workshop agendas for a specific scenario when starting from a blank page

Key Takeaways

Jose’s five points from experience:

  1. Be aware of what you share. Every prompt and attachment you send to a GenAI tool is processed by that service. Be compliant with your customer’s data handling requirements before uploading transcripts, templates, or production data.
  2. More is more. A longer, contextual prompt produces better results than a short, direct one. Include your role, the customer context, the industry, the expected output format — AI cannot infer what is not written.
  3. Always challenge the output. Treat AI answers the way you would treat an unofficial blog post: useful as a starting point, but verify before presenting to a customer or developer.
  4. Stay the owner. The functional consultant owns the FDD, the requirements list, and the test cases — regardless of how much of the content AI generated. Do not attribute errors to AI; review and approve what you deliver.
  5. Define a plan to move up one level. Jose closes with the self-assessment challenge from a Microsoft keynote he attended: rate your current GenAI use (never tried / end user only when required / frequent end user / building own solutions), then make a plan to reach the next level within one year.

Q&A Highlights

One attendee asked what to do when customer processes are undocumented or messy and AI may be working from stale data. Jose recommends using the Microsoft Business Process Catalog as a starting reference — it covers 700+ standard processes across Dynamics 365 and Business Central — and feeding AI the delta between the customer’s actual process and that leading practice. The functional consultant’s job is still to capture and describe the customer’s reality; AI helps once that description exists.


This post was drafted with AI assistance based on the webinar transcript and video content.