Artificial Idea | AI careers · practical prompts · no hype Thursday, November 13, 2025 · Issue #30 · Prompt Tutorial

The data barrier, removed

The data analysis prompt: how non-technical people can make sense of spreadsheets with AI

Every professional is drowning in data and starving for insight. The gap between those two conditions is not a statistics problem. It is a prompting problem.

Issue #29 made the case that the AI transition is producing different opportunities and challenges at different career stages, and that the professionals navigating it best are those who engage with it honestly and specifically rather than through generic upskilling advice. Across every career stage and every story in that issue, one capability appeared consistently as a differentiator: the ability to work with data well enough to make decisions grounded in evidence rather than intuition alone.

This is the capability with the steepest perceived barrier among non-technical professionals and the lowest actual barrier once AI tools are applied to it correctly. The perceived barrier is a statistics and programming barrier. The actual barrier is a question-formation barrier. The professionals who work with data effectively are not primarily those who can run a regression or write a Python script. They are those who can identify the right question to ask of a dataset, evaluate whether the answer they receive is reliable, and communicate what the data shows to an audience that did not see the analysis.

All three of those capabilities are developable with the prompts in this issue, without a statistics course, without a data science background, and without anything beyond access to a spreadsheet and a conversational AI tool.

The question formation problem

Before the prompts, the foundational concept that makes data analysis accessible to non-technical professionals is worth establishing clearly, because it changes how the prompts are used and why they work.

Most non-technical professionals approach data with a question they cannot quite formulate. They have a spreadsheet, they have a sense that it contains something relevant to a decision they are trying to make, and they lack the technical vocabulary to ask for what they need in a way that produces it. The result is either that they do not engage with the data at all, outsourcing the analysis to someone technical and receiving conclusions they cannot evaluate, or that they produce superficial analysis, the sum and the average, that answers no meaningful question about the decision they are trying to make.

The gap is not technical. It is interrogative. The non-technical professional does not need to learn how to analyse data. They need to learn how to ask the right questions of it, how to evaluate whether the answers they receive are reliable, and how to communicate those answers to others. AI handles the technical execution of the analysis. The professional handles the intelligence that makes the analysis relevant to the actual decision.

That division of labour is the one these prompts operationalise.

Prompt 1: The dataset orienteer

The problem it solves: understanding what a dataset contains, what questions it can and cannot answer, and what its limitations are before attempting any analysis, which is the step most non-technical professionals skip and the one most responsible for analysis that misleads rather than informs.

You are a senior data analyst helping a 
non-technical professional understand 
a dataset before conducting any analysis.

Here is a description of my dataset: 
[describe the data: what it contains, 
where it came from, what each column 
or field represents, how many rows 
or records it contains, and the 
time period it covers]

Or here is a sample of the data itself: 
[paste the first 20-30 rows if the 
dataset is not confidential]

The decision or question I am trying 
to answer with this data: 
[describe the business or professional 
question as specifically as possible]

Please:

1. Assess whether this dataset is 
   capable of answering my question 
   as stated, and if not, what 
   additional data would be required
2. Identify the three most important 
   things to understand about this 
   dataset before conducting analysis: 
   its limitations, its likely sources 
   of error, and any structural 
   characteristics that could 
   mislead an inexperienced analyst
3. Reformulate my question into 
   the two or three specific, 
   measurable sub-questions that 
   the dataset can actually answer 
   and that together address my 
   original question
4. Identify the analysis most likely 
   to produce a misleading result 
   with this data if conducted 
   without appropriate caution, 
   and what the caution requires

Do not conduct any analysis yet. 
Help me understand what I am 
working with before I ask 
anything of it.

The instruction not to conduct any analysis yet is the constraint that makes this prompt serve the question formation problem rather than bypassing it. Non-technical professionals who jump directly to analysis without orientation consistently make the same errors: they ask questions the data cannot answer, they do not know when results are misleading, and they cannot evaluate the reliability of conclusions because they do not understand the dataset's limitations. Orientation before analysis prevents all three.

Prompt 2: The pattern finder

The problem it solves: identifying the most significant patterns in a dataset without needing to know in advance what those patterns are, which is the analytical capability most useful for exploratory work where the decision-relevant insight is not yet known.

You are a data analyst conducting 
exploratory analysis to surface 
the most decision-relevant patterns 
in a dataset.

Dataset description or sample: 
[paste or describe]

Business context: [what organisation 
or function this data comes from, 
what decisions are being made 
in this context, and what would 
count as a useful finding]

Please:

1. Identify the three most significant 
   patterns in this data, described 
   in plain language that a 
   non-technical decision-maker 
   can act on, not in statistical terms
2. For each pattern, assess its 
   reliability: is it a robust signal 
   or a pattern that could be an 
   artefact of the data's limitations 
   or a statistical coincidence
3. Identify any pattern that appears 
   counterintuitive relative to 
   what someone familiar with 
   this business context would expect, 
   and flag it for closer examination
4. Identify what the data does not 
   show that would be important 
   for decision-making in this context, 
   the absence of a pattern being 
   as informative as its presence

For each finding, tell me what 
additional analysis would confirm 
or disconfirm it before I act on it.

Point four, what the data does not show, is the analytical observation most consistently absent from data analysis produced by non-technical professionals and most consistently present in analysis produced by experienced ones. Absence of evidence is evidence of absence only under specific conditions, and identifying those conditions requires understanding the data well enough to know what it would show if the pattern existed. This prompt builds that orientation into the analysis.

Prompt 3: The comparison builder

The problem it solves: conducting meaningful comparisons between groups, time periods, or categories in a dataset without falling into the comparison errors that produce misleading conclusions, which are the most common analytical errors in non-technical data work.

You are helping me conduct a comparison 
analysis that will be used to support 
a professional decision.

The comparison I want to make: 
[describe exactly what you want 
to compare: two time periods, 
two customer segments, two products, 
two teams, or any other comparison]

The dataset: [describe or paste]

The decision this comparison will inform: 
[be specific about what you will 
do differently based on the result]

Please:

1. Conduct the comparison I have 
   described and state the result 
   in plain language
2. Assess whether the comparison 
   is a fair one given the data: 
   are the groups being compared 
   similar enough in the relevant 
   dimensions that the comparison 
   is meaningful, or are there 
   confounding differences that 
   could explain the result 
   independently of what I am 
   trying to measure
3. Identify the most likely 
   alternative explanation for 
   the result I have found, 
   assuming the result is real
4. Tell me what sample size 
   or data volume considerations 
   affect the reliability of 
   this comparison, in plain 
   terms rather than statistical ones
5. State whether the result 
   is strong enough to act on 
   or whether it is directional 
   evidence that requires 
   further confirmation

Be direct about the difference 
between a result that is 
statistically reliable and one 
that is suggestive but uncertain. 
Non-technical decision-makers 
are best served by honest 
uncertainty quantification, 
not by false precision.

The instruction to be direct about the difference between reliable and suggestive results is the constraint that makes data analysis honest rather than persuasive. Non-technical professionals frequently receive data analysis that presents uncertain findings with the confidence of established ones, because the analyst is incentivised to produce a clean answer rather than an honest one. This prompt builds the honesty into the output rather than leaving it to the analyst's discretion.

Prompt 4: The insight communicator

The problem it solves: translating the findings from a data analysis into a communication that a non-technical audience can understand, evaluate, and act on, without losing the nuance that makes the analysis reliable.

You are helping me communicate 
data analysis findings to an 
audience that did not conduct 
the analysis and is not 
technically sophisticated.

The analysis I conducted and 
its key findings: [describe 
the analysis and paste or 
summarise the results]

My audience: [their role, 
what decisions they are making, 
what they already believe 
about this topic, and how 
much time they will give 
this communication]

The action I want them to take 
or the decision I want them 
to make based on this data: 
[be specific]

Please produce:

1. A headline finding in one 
   sentence that is accurate, 
   specific, and actionable, 
   not a description of 
   what the analysis did 
   but a statement of 
   what it found
2. Three supporting points 
   that substantiate the 
   headline finding, each 
   in two sentences, 
   each grounded in 
   a specific number 
   from the analysis
3. The most important 
   caveat or limitation 
   that the audience 
   needs to know before 
   acting on this finding, 
   stated in plain language
4. The recommended action, 
   stated specifically 
   enough that the audience 
   cannot misinterpret 
   what they are being 
   asked to do

Constraint: every number 
cited must come from 
the analysis I have described. 
No illustrative figures, 
no rounded approximations 
presented as precise, 
no numbers that cannot 
be traced back to the data.

The constraint that every number must be traceable to the actual analysis is the one that protects the communicator's credibility. Data communications that include illustrative or approximate figures, presented as if they come from the analysis, are one of the most common sources of credibility damage in professional settings. The constraint prevents it.

Prompt 5: The decision validator

The problem it solves: checking whether a decision that is about to be made is actually supported by the data cited in its support, which is the analytical step most frequently skipped and most consequential when skipped.

You are a rigorous analytical reviewer 
checking whether a decision is 
adequately supported by the data 
cited in its favour.

The decision being made: 
[describe specifically]

The data cited in support of it: 
[describe or paste the analysis 
or findings being used to 
justify the decision]

Please:

1. Assess whether the data cited 
   actually supports the specific 
   decision being made, or whether 
   it supports a related but 
   different conclusion that 
   is being extended beyond 
   what the data warrants
2. Identify the gap between 
   what the data shows and 
   what the decision requires 
   to be well-founded, 
   stated as specifically 
   as possible
3. Identify what additional 
   data or analysis would 
   be required to make 
   this decision well-founded, 
   and whether obtaining 
   it is feasible given 
   the decision timeline
4. Give your assessment of 
   whether this decision 
   should proceed as proposed, 
   be modified based on 
   the data limitations, 
   or be deferred pending 
   better information

Be direct. A decision made 
on inadequate data that 
is presented as data-driven 
is worse than a decision 
made on judgment that 
is presented as such, 
because it has false 
confidence attached to it.

The closing instruction reflects one of the most important principles in evidence-based decision-making and one of the least frequently applied. Decisions made on genuine judgment, acknowledged as such, can be evaluated and revised when the judgment proves wrong. Decisions made on inadequate data, presented as data-driven, are defended with the authority of evidence they do not actually have, which makes them harder to revise when they prove wrong and more damaging when they are.

The thirty-minute data practice

These five prompts used in sequence turn any dataset into a decision-relevant analysis in under thirty minutes for a professional who has no technical background and has never written a line of code.

The sequence: orient before analysing, find patterns before testing hypotheses, build comparisons carefully, communicate findings honestly, and validate that the decision is actually supported by the data before making it.

That sequence is what experienced data analysts do. It is what separates analysis that informs decisions from analysis that decorates them. The technical execution is now available to anyone with a prompt and a dataset. The intelligence that makes the execution relevant to the actual decision has always been the scarce resource, and it remains the scarce resource.

These prompts put the technical execution within reach. The intelligence is yours to supply.

Monday we are examining one of the most consequential and least discussed dynamics in the AI and careers space: the specific moment at which the accumulation of small, consistent investments in AI capability begins to produce returns that are visible to people other than the investor. The compounding curve has an inflection point, and understanding where it is changes how you think about the investments that precede it.

The inflection point is closer than most professionals currently believe. Monday explains why.

— The Artificial Idea team

Keep Reading