Support page

Sources and methodology

This page explains how the Animals in Disasters evidence map was built, what it includes, what it does not claim to be, and where the current event-level source material comes from. It is designed to support transparency, defensibility and internal review.

20 documented map events · per-event source links · confidence labels included

Why this page exists

The map is designed to show a policy pattern, not to function as a complete national casualty database.

The Crisis page uses a selective set of documented disaster events to show that animals are repeatedly affected across fire, flood, storm, drought, heat and marine events — and that animal welfare planning remains fragmented before, during and after emergencies.

What this page supports

It gives internal and external readers a transparent view of the evidence architecture behind the map: why events were chosen, how source confidence is handled, and how to verify the public sources currently attached to each event card.

Why that matters

Because the map is a public-facing advocacy tool, it needs to be credible, appropriately bounded, and open about its limitations. This page helps prevent the map from being read as a claim to total coverage or perfect quantification across all disasters.

How the map was built

Five principles underpin the current evidence sample.

The map is curated around events where animal impacts were documented clearly enough to support the campaign’s policy argument, while still showing variation across disaster type, geography, timeframe and affected animal groups.

01 · Inclusion

Events were selected for documented animal relevance.

The map prioritises events where animal impacts, rescue burdens, habitat loss, mortality, or animal-welfare failures were described strongly enough in public material to support inclusion.

02 · Breadth

The sample spans more than one disaster type.

The index includes fire, flood, drought, heat, cyclone, storm, marine and freshwater events so the campaign does not imply that animal disaster planning is only a bushfire issue.

03 · Traceability

Every current event card carries a public source link.

Per-event links were added during the source-review pass so readers can inspect the underlying material used for each mapped event and see how the current card framing was derived.

04 · Revision

Claims were tightened where stronger public evidence existed.

Some earlier figures and phrasings were revised during QA to avoid overstating uncertain totals and to replace weaker claims with more defensible event-specific framing.

05 · Representation

Coordinates are indicative event locations, not disaster boundaries.

Map points are used as orientation aids. They do not attempt to represent the full footprint of each disaster or the full spatial distribution of animal impacts.

Scope and limits

What the map does — and what it does not claim to do.

Keeping these boundaries clear is important for credibility. The map is intended to help readers understand a recurring policy problem, not to imply exhaustive national event coverage or uniformly quantified animal tolls.

What it includes

  • A curated cross-section of Australian disasters affecting wildlife, farmed animals, companion animals or rescue systems.
  • Events where the public source trail is strong enough to support campaign framing and external review.
  • Historical and recent examples to show recurrence, policy gaps and the need for pre-disaster planning.
  • Event cards that prioritise legibility and accountability over exhaustive incident detail.

What it does not do

  • Claim to list every Australian disaster affecting animals.
  • Claim a fully quantified national dataset of animal casualties across all event types.
  • Use point markers as exact impact boundaries or precise event footprints.
  • Treat all source types as equal in confidence, completeness or methodological strength.
Source confidence

The map now distinguishes stronger evidence from contextual or partial material.

Confidence labels are designed to signal how firmly the current event framing rests on the available public source. They are not a formal peer-review system, but a practical campaign QA layer.

High

Government, scientific, or strongly attributable public material

Used where public reporting, scientific reviews, inquiries, government documents or other authoritative records support the current event framing with strong traceability.

Medium / Medium-high

Useful public reporting, but with more caution around precision

Used where the event is clearly relevant but the public record is less complete, more media-dependent, or more preliminary than the strongest sources in the map.

Context only

Retained mainly to show historical pattern or known data limits

Used where an event remains important to the narrative but the available public source material does not justify strong quantification or should be treated primarily as context.

Event source index

Current map events and their public source trail.

This index reflects the current reviewed map dataset. It is sorted from most recent to oldest so additions and revisions are easier to inspect. Each card links directly to the source currently attached to the corresponding map event.

20 map events Recent and historical examples Wildlife, farmed, marine and mixed impacts Public links for external review

Showing 6 of 20 mapped events. Expand the index to inspect the rest of the current public source trail.

Update and revision approach

The map should remain curated, reviewable and open to correction.

Future changes should strengthen credibility rather than expand the map indiscriminately. The goal is not to create a noisy database, but a defensible evidence tool that supports the campaign’s case for a Disaster Plan for Animal Welfare.

Additions should fill real evidence gaps.

New events should be added when they materially improve disaster-type coverage, species-group coverage, recency, recurrence, or policy relevance — not just because another headline exists.

Claims should tighten when stronger sources emerge.

If a better government, scientific or inquiry source becomes available, event language should be revised to match it, even if that means reducing rhetorical force in favour of defensibility.

Context should stay visible.

Historical events and unquantified impacts can still be worth including, provided they are clearly labelled and not presented as stronger or more precise than the evidence allows.

Back to the campaign

Evidence matters. So does what we do with it.

The map’s role is to show a recurring pattern: animals are repeatedly affected in disasters, but planning remains fragmented. From here, the next steps are to review the evidence in context, read the proposed framework, and help build pressure for change.