Download Effective Deployment and Communication

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Internal communications wikipedia , lookup

Business model wikipedia , lookup

Transcript
Session One – Advancement Data:
Metrics and Communication
Your data may be complete and thoroughly clean;
your metrics may be perfectly designed to measure
what matters; yet without a thoughtful approach to
getting that information into the right hands at the
right time with the right context, your efforts may
still fall short. This session will provide an
organizational framework and proven techniques for
success.
Advancement Data:
Metrics and Communication
Lisette Clem ‘85 ‘92MBA
Director of Advancement Services
Bryant University
Smithfield, RI
Agenda
 Define “internal constituents”
 Why share information?
 What advancement data/metrics do we share?
 When do we share?
 How is the information shared (in what format)?
Internal Constituents
 Outside of the Advancement division (be sure to
include those folks with whom you share
external constituents!):
 Controller’s Office
 President’s Office
 Other organizational divisions to whom donations are
being directed
Internal Constituents
 Within the Advancement division:
 Alumni/Constituent Relations
 Development
 Marketing/Communications
 Within Advancement Services:
 aka “the Cool group”
 aka “Team Awesome”
TODAY’S FOCUS: Using the Power of Information
Sharing in engaging and informing our divisional
constituents (and getting them to pay attention!)
Our Goal:
To CONSISTENTLY provide a Proactive vs. Reactive
approach to information sharing and analysis, focusing on
both detailed reporting (for Annual Fund managers) and
high-level summaries (for VPs) to accommodate everyone’s
needs*
*In this case study: without the benefit of a Business Intelligence
system or built-in dashboards (it’s coming!)…
WHY should we
proactively share
information/data?
(What does it matter?)
Why?
 Communication and analysis of the usefulness of
the data in fact justifies renewed investments in
technology
 Fewer data requests in to our Report Writing
staff; enables multi-tasking
 Enhanced stewardship (for soft credit gifts)
 More effective prospect management (air of
friendly competition at monthly PM meetings)
 Less chaos prior to Trustees’ meetings (!)
WHAT do we share?
 General Development Performance Reporting:
 Gift and Pledge Processing (daily transaction reports)
 Fiscal Year Status
 Campaign Reporting
 Prospect Management
 Annual Giving
 Fiscal Year performance and Trend data
 Alumni (Constituent) Participation Rate
 Data Mining model performance
 Definition/Results
 Strategy Recommendations
 Event Management
 Budget Reporting
 Expenses/Revenue
 Return on Investment
 Alumni Engagement Tracking
Campaign
Reporting
Prospect
Management
Prospect Management:
Key areas of reporting
1.
Prospect Pool
2.
Visit Reports
3.
Tickler Reports (to support Moves Management)
4.
Results Reports
 Pledges and Gifts
 Planned and Pending Solicitations
5.
Ongoing Program Management
Visit Report
Visits in FY14 vs. FY13 (as of X date)
Summary Results Report – Part I
Summary Results Report – Part II
Ongoing Program Management:
The “RED” Report
Annual Giving:
Fiscal Year performance
and Trend data
Fiscal Year Annual Giving
Fiscal Year Annual Giving
Fiscal Year Annual Giving
FY14 vs. FY15 New Pledges
Monthly Comparison
(Excludes Planned Gifts)
As of 5/31/15
Constituent Participation Rate
 The Politics of Participation
 One size does NOT fit all
 CASE , VSE, and US News have standards for their
comparisons


These standards may or may not be useful, or even accurate,
for your purposes
When submitting for them, make every attempt to
understand their standards, to preserve the value of
benchmarking with other institutions
Alumni Participation Rate
(at Bryant University)
 In both our external and our internal alumni participation rates, senior
giving donors are counted as alumni donors, as anyone who attends Bryant
for 2+ semesters has the option of calling themselves an alumnus/ae. The
entire senior class is included in the denominator.
 Our “lost” alumni percentage is approximately 5.5%, down from
approximately 10% four years ago. ALL NOT-LOST Alumni are
considered “of record” per CASE standards, regardless of giving history.
 In our internal donor count, all soft credit donors ARE included; however,
approx. half of these are alum/alum spouses, which CAN be counted per
CASE. (Counting the remainder of our soft credit donors accounts for
approximately +0.4% of our internal participation rate.)
 Our internal alumni participation rate excludes those on a “Do Not Solicit
(DNS)” code (approx. 650 DNS alums of our 36,500 alumni of record).
 Including ALL degree holders vs. only including UG degree holders
decreases our alumni participation rate by just 0.2%.
Data Mining:
Definitions/Results and
Strategy Recommendations
Current Models
 Affinity Insight Acquisition Scoring Model: The scores from this
model can be used to identify those never donors most likely to be
responsive to annual fund appeals.
 Predictive Affinity Retention Scoring Model: The scores from this
model can be used to identify and retain those alums who have
given in the past but have not made a gift in the current fiscal year.
 Predictive Affinity Donor Scoring Model: The scores from this model
can be used to identify alums most likely to make “leadership-level”
annual gifts, as well as new major giving prospects.
FY12 Results – Acquisition Model
Data Mining Acquisition Model
New donors in FY12
Scores of 1-16
48%
Scores of 17-20
52%
(Recall that each score, or decile, represents 5% of our alumni
population, so expected results would be 20% for four scores.)
FY12 Results – Donor Scoring Model
Data Mining Donor Scoring Model
Average Gift Size
$646
Scores of 1-16
$800
Scores of 17-20
$0
$100
$200
$300
$400
$500
$600
$700
$800
Data Mining Strategy:
Recommendations
 Acquisition Model:
 “Acquisition Campaign” – New Donor Drive



Targeted DM to scores 15-20 – “flyer” format (not letter!)
Telefund focus on scores 15-20
Targeted e-solicit messages to scores 15-20
 Retention Model:
 “Retention Campaign” – Save a Donor


Intense Telefund and E-Solicit focus to scores 15-20
Follow up with personal outreach to all those not yet renewed by 5/31/13 –
Firm Goal is 100% renewal for scores 15-20
 Donor Scoring Model:
 Leadership Giving Campaign – Raising Sights


Assign scores 17-20 to Bryant Fund Leadership Giving officer
Personal Outreach
Data Mining: Coming Soon!!
 Predictive Affinity Discovery Scoring Model:
This model will focus on those alums who are most
likely to accept a Discovery Visit invitation. The model
will be built from alums who have been asked to accept
a Discovery Visit. The focus of the model will be on
alumni database variables that discriminate between
those alums who accepted a Discovery Visit and those
who didn’t. The result of the modeling process will be a
score for each reachable alum in the database.
Event
Management
Budget Reporting:
Expenses/Revenue and
Return on Investment (ROI)
Cost per dollar raised:
Revenue basis
Benchmarking: ROI
Alumni Engagement
Tracking
Framework for Alumni Engagement - Tracking - updated August 2013
1
2
3
Category
Description
Giving
Gifts or pledge payments (hard/soft) from
alumni for any purpose (within the current
fiscal year)
Cash: hard/soft by fiscal year
Total FY giving (hard/soft) > $0 (not to include
SCG).
Feedback/ Feature
Direct feedback offered to the University
regarding a communication, alumnus/a
featured in University publication or
communication, or alumnus/a featured in
web content.
Communication from alum, either via in-person
conversation, phone converation or email
offering feedback re: a University publication or
web posting. Also, alumnus/a serving as a
subject of a profile in a publication or via web
content.
Contact report with type: FEA. Also, entry in
Comments with a subject code of XALPRO or
CHAIR.
Attendance, participation or hosting of
events on and off campus.
Event attendee code (excluding SAA dinner and
Alumni/Student networking dinners); also,
contact code indicating that an alum was a
featured speaker or invitee to a campus event.
Event attendee code on APASPUR, date driven.
(Some exclusions apply.) Also, contact code of
IPS or HST.
Activity codes and event attendee codes.
APAACTY codes associated with Alumni
Admission Associates (year indicator required),
Career Link (VCareer - any year), and/or Career
Shadowing program (year indicator required);
APASPUR codes associated with attendees to
Alumni/Student networking events and the SAA
dinner.
Events
Source of data
Reporting criteria
Alumni/ Student Engagement
Alumni/student related activities including
Alumni Admission Associates, Career
Center volunteers, and others.
5
Alumni Leadership Activities
Traditional alumni volunteer roles
including NAC and RAN groups, as well as
class steering committees, alumni trustees,
Senior Advisory Committee members,
trustees, etc.
Activity codes.
APAACTY codes for NAC, RAN, Dean's Advisory
Committee, Class Steering committees, trustees,
and immediate past Senior Advisory Committee
members. NOTE: Activity codes for these
activities MUST be consistently coded by fiscal
year.
6
Prospect Visits
Alumnus/a participated in a Discovery
and/or Cultivation visit(s)
Visit codes.
Contact codes of VDI, VSO, VST and/or VCU.
7
BLC Member Y/N
BLC ($1,000+) member current fiscal year
(Y/N)
Gift Society member status
Gift Society level current FY at BLC level +.
8
Linked-In and/or OLC
Member Y/N
Participation in Bryant's Linked-In
community or the iModules online
community
Activity code indicating status as a member of
Bryant's Linked-In community or the iModules
OLC
APAACTY code associated with Linked In
("LINKED") and APAACTY code indicating OLC
member status. Year indicators not needed.
Is an alum parent with a child who either
is currently matriculating at Bryant or is
an alum of Bryant
Constituent codes
APACONS codes of PART/PARF AND
ALUG/ALND/ALUG.
4
9
Legacy
WHEN do we share?
“when it happens”
Daily
Weekly
Monthly
Quarterly
Annually
HOW do we share the information?
Paper
Email
(Internal) Web site
Regular Meeting
Common drive
Other?
(Soon: Dashboard!)
Other Useful Examples:
 Advancement Services Information, Reference
and Forms Guide
 Advancement Services Information and
Resources Web page (live quick links)
 Advancement Services Annual Report
 Fails 
 TOO MANY prospect management reports –
negative impact on perceived relevance
 Don’t save copies of everything (maxed out our
server allotment!)
 Edits/customization by person will be requested –
saying no is OK! (unless it’s the VP)
Other suggestions/ideas,
management tips, best practices -what works for you to keep
colleagues, constituents and staff
informed?
Thank you!
Have a great day!

Lisette Clem, Bryant University
[email protected]
401-232-6805