🧭 Overview / What This Guide Covers
An investment screening checklist is the fastest way to turn messy inbound deals into consistent decisions. This guide shows you how to build a one-page template that standardises your investment screening method, reduces bias, and flags issues early-before you spend time on deep modelling or diligence. It’s built for CFOs, corporate development, investors, and analysts who need repeatable investment evaluation across many opportunities. You’ll leave with a checklist you can reuse, score, and share-aligned to your broader investment screening process-so every deal gets the same baseline review and a clear next step.
✅ Before You Begin
Before you begin, pick the context your checklist must serve: investment opportunity screening for new deals, project investment screening for internal capex, or ongoing investment project evaluation for portfolio add-ons. Gather (1) your decision gates (who approves, what needs escalation, what is “auto-no”), (2) required return metrics (IRR, NPV, payback) and the minimum inputs you can reliably get at intake, (3) strategic constraints (industry, geography, ticket size, leverage limits), and (4) a short list of non-negotiables (regulatory, reputational, concentration risk). If your team already uses scoring, confirm the weightings so your checklist matches your strategic investment screening approach. You’re ready when you can answer: what does “good” look like, what can be deferred to diligence, and what data you will not chase without a pass on the initial screen. If you plan to model quickly, decide the base-case structure for your investment screening model (time horizon, discount-rate convention, and unit economics level).
🛠️ Step-by-Step Instructions
Step 1: Define the decision gate and screening scope
Start by writing the decision gate at the top of the page: “Proceed to diligence”, “Request more info”, or “Decline”. Then define the scope of your investment screening: what you will evaluate at intake vs what happens later. One of the simplest investment screening steps is to separate “must know” from “nice to know”. Must-know items typically include the problem being solved, how money is made, basic traction, and headline economics; nice-to-know items include detailed cohort tables, legal structure nuance, or integration roadmaps. Add a short “deal snapshot” block (company/project, size, use of funds, timeline) and a “fit” block (strategy, customer, channel, capability). This keeps your investment analysis consistent and prevents teams from over-indexing on a great story with weak fundamentals. If you want the checklist to map cleanly into an end-to-end investment screening process, assign a single owner per block and timebox the first pass.
Step 2: Build the checklist sections that drive fast yes/no decisions
Now create 6–8 sections with short prompts (not long questions). A strong default is: (1) Strategic fit, (2) Market & competition, (3) Product/operations, (4) Team & governance, (5) Financials & returns, (6) Risks & deal terms. Under “Financials”, focus on financial investment screening: what minimal numbers can you validate today (revenue model, gross margin, retention, capex intensity, working capital drag)? Add a tight “returns box” with the three numbers decision-makers read first-IRR, NPV, and payback-plus the single assumption behind each (volume, price, margin). This keeps investment evaluation grounded in cash, not narratives. If your organisation uses formal project investment appraisal, add a checkbox row for benefit category (growth, cost, risk, compliance) and who owns delivery. For selection logic behind NPV vs IRR vs payback, keep a simple reference point so reviewers don’t argue methods instead of facts.
Step 3: Add scoring, thresholds, and red-flag logic
A checklist becomes powerful when it forces decisions. For each section, add a 1–5 score with a one-line definition of what “1” and “5” mean. Then set thresholds: e.g., any single “1” in regulatory risk triggers “Decline”, while an average score below 3 triggers “Request more info”. This is where investment risk screening lives-your early warning system for leverage traps, customer concentration, fragile unit economics, or unfundable capex. Make red flags explicit (e.g., “gross margin < X% with no path to improvement”, “one customer > Y%”, “negative working capital turns driven by unsustainable terms”). Include a small “evidence” column so every claim links to a source (data room doc, call note, financial statement). That one column turns the page into auditable investment project evaluation rather than opinion. For a deeper list of risk signals to include, borrow the structure from an investment risk screening guide and adapt it to your sector.
Step 4: Connect the checklist to a lightweight model and sensitivities
At the end of the page, add a “Model in 10 minutes” mini-block: 5–8 drivers that explain 80% of the value (price, volume, conversion, churn, gross margin, capex, working capital days, cost-to-serve). The goal isn’t precision; it’s to translate your checklist into a defensible investment screening model you can stress-test. Document the base-case assumptions as ranges, not single points, then run a quick one-way sensitivity on the top two drivers. This links qualitative investment screening to quantitative investment analysis without building a full diligence model. If your team uses Model Reef, this step becomes faster because you can create driver-based structures once and reuse them across deals, including consistent cash flow and valuation outputs. Keep the checklist and the model aligned: every scored item should map to one driver or one risk assumption, so your investment screening method stays coherent from intake to recommendation.
Step 5: Operationalise the checklist for reuse, collaboration, and governance
Finally, make the checklist usable at scale. Turn it into a reusable template (one page, same order, same scoring). Define roles: who completes the first pass, who reviews, and who makes the call. Add a timebox (e.g., 24–48 hours from intake to decision) to protect focus. For shared pipelines, create a naming convention so you can compare investment opportunity screening outcomes across deals (e.g., “Sector – Company – Stage – Status”). Store the checklist next to your model and call notes so reviewers can trace how the decision was made. In Model Reef, teams can keep the underlying assumptions, scenarios, and commentary in one place, making review and handoffs clean-especially when multiple stakeholders need to sign off. The output should be a single, clear decision plus the next action: decline, diligence, or revise the case.
⚠️ Tips, Edge Cases & Gotchas
Treat the checklist as a guardrail, not a straightjacket. Early-stage deals often fail traditional financial investment screening because revenue is nascent; in that case, shift weight to market, unit economics logic, and execution capacity-but keep the same investment screening steps so decisions remain comparable. For internal capex, don’t copy a venture checklist: project investment screening needs explicit implementation risk, downtime impact, and who owns benefits realisation. Watch for double-counting: if you score “market size” and also bake aggressive growth into the model, you can inflate returns twice. Another common pitfall is scoring without evidence-require a citation for each key claim (even if it’s just a call note). If you need an audit trail, use a workflow with built-in change tracking, comments, and version history so later reviewers can see what changed and why. Finally, define one exception path (e.g., strategic adjacency bets) with a separate label and additional approval, so your core investment screening method doesn’t get diluted by one-off pet projects.
🧪 Example / Quick Illustration
You receive a $5M growth capex proposal to automate a production line. Intake data: $1.2M annual labour savings, $0.3M annual maintenance increase, $5M capex split 60/40 over 6 months, 8-year life, and a 10% discount rate. Action: run the checklist. Strategic fit scores 4/5; implementation risk scores 2/5 due to vendor dependency (flag). Build a mini investment screening model with drivers: savings ramp, downtime weeks, capex timing, and working capital impact. Then run two sensitivities: savings ramp (6–18 months) and downtime (0–6 weeks). Output: base case NPV is positive and payback is 3.2 years, but the downside case turns negative if downtime exceeds 4 weeks. In Model Reef,you can set those ranges as scenarios and instantly compare base vs downside outputs, making the “proceed to diligence” decision defensible.
🚀 Next Steps
Next, convert your checklist into a repeatable workflow: define your intake SLA, assign owners, and decide how “proceed to diligence” gets triggered. Then audit the first 10 screenings to see which questions actually changed decisions-and cut the rest. If you want to scale investment screening across a pipeline without spreadsheet sprawl, Model Reef can help you standardise drivers, scenarios, and approvals so each opportunity is evaluated the same way,with clear traceability from assumptions to outputs. Once your checklist is live, your next upgrade is to tighten scoring and link it to a lightweight model so the team can defend outcomes under challenge.