Business Intelligence Foundations: What BI Delivers and How It’s Used

Capítulo 1

Estimated reading time: 8 minutes

+ Exercise

Business Intelligence (BI) is the practice of turning data into trusted, decision-ready information. In practice, BI connects raw operational data (often messy, inconsistent, and spread across systems) to the questions people ask at work: “Are we on track?”, “What changed?”, “Where should we act?”, and “Did it work?”

BI is not just “making charts.” It is a set of methods and products that ensure the same metric means the same thing across teams, that numbers can be traced back to sources, and that insights arrive at the right time and level of detail to support decisions.

What BI delivers: decision-ready outputs

BI outputs are designed to be consumed repeatedly and reliably. The same dataset can produce multiple outputs depending on the audience and use case.

OutputWhat it isBest forTypical cadence
DashboardsInteractive visual views of key metrics and trendsMonitoring and explorationNear-real-time to daily/weekly
ReportsStructured, often paginated tables with filters and totalsOperational and compliance-style needsDaily/weekly/monthly
KPI scorecardsSmall set of agreed KPIs with targets, status, and ownershipPerformance managementWeekly/monthly/quarterly
Ad hoc queriesOne-off questions answered by slicing dataInvestigations and hypothesis checksOn demand

To be “decision-ready,” these outputs typically include: consistent definitions (e.g., what counts as “active customer”), clear time windows (e.g., last 7 complete days), and enough context to act (e.g., breakdown by region, product, channel).

Common BI use cases (and how they differ)

1) Performance monitoring

Goal: Detect whether the business is on track and spot deviations early.

Continue in our app.
  • Listen to the audio with the screen off.
  • Earn a certificate upon completion.
  • Over 5000 courses for you to explore!
Or continue reading below...
Download App

Download the app

Typical outputs: KPI scorecards and dashboards with trends, targets, and alerts.

Common questions:

  • Are we hitting revenue and margin targets this week?
  • Is conversion rate dropping in any channel?
  • Which regions are behind plan?

Practical example: A weekly sales scorecard shows Revenue, Gross Margin %, New Customers, each with a target and a red/yellow/green status. A drill-down dashboard lets managers click “Revenue” to see it by region → store → product category.

2) Operational reporting

Goal: Run day-to-day operations with accurate, detailed lists and totals.

Typical outputs: Paginated reports, scheduled email reports, exception lists.

Common questions:

  • Which orders are delayed beyond SLA?
  • Which invoices are overdue and by how many days?
  • What is today’s inventory by warehouse and SKU?

Practical example: A daily “Late Shipments” report lists each shipment with order ID, promised date, carrier, warehouse, and delay reason. The report is filtered to “only exceptions” so teams focus on what needs action.

3) Exploratory analysis

Goal: Understand “why” something happened and identify drivers and patterns.

Typical outputs: Interactive dashboards, ad hoc queries, exploratory notebooks (depending on the organization), and saved analysis views.

Common questions:

  • Why did churn increase last month?
  • Which customer segments are most sensitive to price changes?
  • What factors correlate with support ticket volume?

Practical example: An analyst starts from a churn dashboard, filters to the affected month, segments by plan type, and then checks whether churn is concentrated in customers with high ticket counts and recent billing issues.

4) Self-service analytics

Goal: Enable non-technical users to answer their own questions safely, without reinventing metrics.

Typical outputs: Curated datasets, governed semantic models, certified dashboards, and ad hoc query interfaces.

Common questions:

  • How many leads did my campaign generate by week and region?
  • What is the average time-to-resolution for my team?
  • Which products are most frequently returned in my category?

Practical example: Marketing has a certified “Campaign Performance” dataset with standardized definitions for Lead, MQL, Attribution Window. A manager can build a simple chart without needing to join tables or interpret raw event logs.

Choosing the right output for the job

Different use cases favor different BI products. A useful way to decide is to map the question to the output type.

If the user needs…Prefer…Because…
Fast status vs targetKPI scorecardMinimizes noise; focuses on accountability
Ongoing monitoring with drill-downDashboardCombines overview + interactive investigation
Row-level lists for actionOperational reportSupports workflows (call, fix, ship, reconcile)
One-off “why/what-if” questionsAd hoc queries / exploratory viewsFlexible slicing without rebuilding a report

Stakeholder needs and how they shape BI requirements

BI succeeds when it matches how different stakeholders make decisions. The same metric (e.g., “on-time delivery”) may need different latency, granularity, and interactivity depending on who is using it.

Executives

Primary need: Directional clarity and accountability across the business.

  • Latency: Often daily/weekly is sufficient, but critical metrics (cash, incidents) may need near-real-time.
  • Granularity: High-level with the ability to drill to business unit/region.
  • Interactivity: Moderate; drill-down and filters, but not complex configuration.

What “good” looks like: A scorecard with a small set of KPIs, clear targets, and consistent definitions across departments.

Managers (functional and mid-level)

Primary need: Diagnose performance and allocate resources.

  • Latency: Daily or intraday for operational areas (support, logistics); weekly for planning cycles.
  • Granularity: Team/product/channel level; needs segmentation and comparisons.
  • Interactivity: High; drill-down, slice-and-dice, cohort views, and the ability to save filtered views.

What “good” looks like: Dashboards that start with KPIs but quickly answer “where is the problem?” and “what changed?”

Frontline teams (operations, sales reps, support agents)

Primary need: Take immediate action on specific items.

  • Latency: Near-real-time to intraday for many workflows.
  • Granularity: Very detailed (row-level), often tied to a case/order/customer.
  • Interactivity: Focused; filters and search are key, plus exports or integrations into workflow tools.

What “good” looks like: Exception reports and work queues (e.g., “tickets breaching SLA in the next 2 hours”).

How latency, granularity, and interactivity trade off

These three requirements often compete. Higher granularity and lower latency can increase cost and complexity, and may reduce performance. BI design typically makes explicit choices:

  • Latency vs trust: Near-real-time data may be less validated; daily data can be more reconciled and consistent.
  • Granularity vs speed: Row-level dashboards can be slow; aggregated views are faster and easier to interpret.
  • Interactivity vs governance: More freedom can increase the risk of misinterpretation unless definitions and guardrails are strong.

Step-by-step: mapping a business question to a BI deliverable

Use this practical sequence to decide what to build and how to build it.

  1. Write the decision and the action. Example: “If on-time delivery drops below 92% in any warehouse, we will add carrier capacity or reroute shipments.”

  2. Choose the primary audience. Example: operations managers (diagnose) and frontline coordinators (fix specific shipments).

  3. Define the metric and its boundaries. Example: On-time delivery % = delivered_on_or_before_promised_date / total_delivered; exclude canceled orders; use promised date from the order system.

  4. Set latency and refresh expectations. Example: refresh every 2 hours for today’s operations; daily reconciled view for weekly performance review.

  5. Pick the output(s). Example: a manager dashboard (trend + breakdown by warehouse/carrier) plus an exception report listing late-risk shipments.

  6. Specify required drill-down and filters. Example: warehouse → carrier → route; filters for date, service level, customer tier.

  7. Define ownership and validation. Example: operations owns the definition; data team validates joins and edge cases; business signs off on KPI logic.

End-to-end BI flow and responsibilities

Most BI implementations follow a flow that turns raw data into decisions. The key is that each step implies responsibilities for quality, clarity, and usability.

Source data → Transformation → Semantic layer → Visualization → Decisions

1) Source data

What happens: Data is generated by operational systems (e.g., sales transactions, support tickets, web events, finance postings).

Responsibilities implied:

  • Identify authoritative sources for each domain (e.g., billing system for revenue recognition fields).
  • Understand data meaning and limitations (timestamps, late-arriving records, missing values).
  • Ensure access and basic security (who can read what).

2) Transformation

What happens: Data is cleaned, standardized, joined, and reshaped into analysis-ready tables (e.g., consistent customer IDs, standardized product categories, derived fields like “week start”).

Responsibilities implied:

  • Implement repeatable logic for cleaning and joining data.
  • Handle edge cases explicitly (refunds, cancellations, duplicates).
  • Validate outputs (row counts, reconciliation totals, anomaly checks).

3) Semantic layer

What happens: Business-friendly definitions are created: metrics, dimensions, hierarchies, and relationships (e.g., “Net Revenue,” “Active Customer,” “Region → Country → City”).

Responsibilities implied:

  • Define metrics once and reuse them everywhere to avoid metric drift.
  • Document definitions and calculation rules.
  • Apply governance: certified datasets, access rules, and consistent naming.

4) Visualization

What happens: Dashboards, reports, and scorecards are built to answer specific questions with appropriate interactivity.

Responsibilities implied:

  • Design for the audience: executive overview vs operational detail.
  • Choose visuals that match the task (trend lines for monitoring, tables for action lists).
  • Ensure usability: clear filters, sensible defaults, performance optimization.

5) Decisions

What happens: People act on the information: adjust budgets, change staffing, fix process issues, prioritize accounts, or investigate anomalies.

Responsibilities implied:

  • Assign KPI owners and define what actions are triggered by thresholds.
  • Close the loop: track whether actions improved outcomes.
  • Maintain feedback channels so BI products evolve with changing needs.

Now answer the exercise about the content:

A team needs a BI deliverable that provides a row-level list of items requiring immediate action (for example, exceptions to fix in daily operations). Which output is the best fit?

You are right! Congratulations, now go to the next page

You missed! Try again.

Operational reports are designed for day-to-day workflows and provide detailed, row-level lists (often filtered to exceptions) so teams can take action on specific items.

Next chapter

Core BI Vocabulary: Facts, Dimensions, Measures, and Grain

Arrow Right Icon
Free Ebook cover Business Intelligence Foundations: Concepts, Roles, and Common Architectures
10%

Business Intelligence Foundations: Concepts, Roles, and Common Architectures

New course

10 pages

Download the app to earn free Certification and listen to the courses in the background, even with the screen off.