Tag Archives: nptech

Nonprofit Radio for April 13, 2026: Responsible AI Adoption & Ethically Using AI

 

Jen Frazier, Kate Dreyfuss & Ben Freda: Responsible AI Adoption

We continue our coverage of the 2026 Nonprofit Technology Conference (26NTC), with a panel that helps you find the low-hanging fruit for AI at your nonprofit. They share their 5-Step framework for deploying AI in small, thoughtful steps. They’re Jen Frazier, from Firefly Partners; Kate Dreyfuss at New Music USA; and, Ben Freda with BFC Digital.

 

Chris Rosica & Bethany Friedlander: Ethically Using AI

Chris Rosica and Bethany Friedlander share 5 strategies to create compelling web and social content with AI, keeping in sight ethics and risks: Ideation; targeting; optimization; repurposing; and, staying human. Chris is from Rosica Communications and Bethany is at New Bridge Cleveland.

 

Listen to the podcast

Get Nonprofit Radio insider alerts

 

Apple Podcast button

 

 

We’re the #1 Podcast for Nonprofits, With 13,000+ Weekly Listeners

Board relations. Fundraising. Volunteer management. Prospect research. Legal compliance. Accounting. Finance. Investments. Donor relations. Public relations. Marketing. Technology. Social media.

Every nonprofit struggles with these issues. Big nonprofits hire experts. The other 95% listen to Tony Martignetti Nonprofit Radio. Trusted experts and leading thinkers join me each week to tackle the tough issues. If you have big dreams but a small budget, you have a home at Tony Martignetti Nonprofit Radio.
View Full Transcript

And welcome to Tony Martignetti Nonprofit Radio. Big nonprofit ideas for the other 95%. I’m your aptly named host, and I’m the pod father of your favorite hebdominal podcast. Oh, I’m glad you’re with us. I’d be forced to endure the pain of ende or Titus if you inflamed me with the idea that you missed this week’s show. Here’s our associate producer, Kate, with what’s up. Hey Tony, I’m on it. Responsible AI adoption. We continue our coverage of the 2026 nonprofit Technology conference with a panel that helps you find the low-hanging fruit for AI at your nonprofit. They share their five-step framework for deploying AI in small, thoughtful steps. They are Jen Frazier from Firefly Partners, Kate Dreyfus at New Music USA and Ben Freda with BFC Digital. Then Ethically using AI. Chris Rossica and Bethany Friedlander share 5 strategies to create compelling web and social content with AI, keeping in sight ethics and risks, ideation, targeting, optimization, repurposing, and staying human. Chris is from Rossica Communications, and Bethany is at New Bridge, Cleveland. On Tony’s take 2. Tales from the gym. Meet chatty Sam. Here is responsible AI adoption. Welcome back to Tony Martignetti nonprofit radio coverage of 26 NTC. You know that that’s the 2026 nonprofit technology conference that all these smart tech folks and and tech users, this is not strictly a conference for tech employees, tech tech professionals. This is also for all professionals who use technology. So unless you’re still using index cards as a CRM, you are using technology. odds are. You’re using Microsoft 360 most likely, so 365, 365 or 360? 365, 365. It’s all it’s every day of the year, Tony, every day. Well, it could have been 360 degrees. I did take 5 days off. All right, yelling. It could have been 360 degrees, degrees of a compass. I was a Boy Scout. Now you’re mocking the Boy Scouts. I am. We haven’t even, we haven’t even introduced, all right, I’m gonna, I’m shutting our mic off. That’s it, we, we now have two panelists instead of three. All right, so we’re at the 2026 nonprofit technology conference in Detroit at Huntington Place. With me now were going to be Jen Frazier. She’s no longer, but she’s she’s no longer a guest, but she’s still CEO and founder of Firefly Partners. Also Kate Dreyfus, director of communications at New Music USA, and Ben Freda, president at BFC Digital, almost like BFD should have been BFD Digital. Big fucking deal. That’s right, VFC, man. Whatever. There’s still time to change it. I mean, he changes that stuff all the time, so, alright, uh, and their topic is low hanging fruit, clearing a path for responsible AI adoption. At your organization I I I trimmed off at your organization. I think that’s understood, but low hanging fruit, clearing a path for responsible AI adoption. um, Jen, you’re you’re the best person I think to give it just an overview of the topic, please, please, yeah, so I think right now obviously lots of folks are excited and a little bit overwhelmed by. Sort of the idea of AI and what can I do with this amazing kind of new technology that has come roaring into our lives and so what we wanna do is just give a quick overview for folks about how to take some of the overwhelm and to really just boil it down to some pretty concrete simple steps to get started so we have a. A nice 5 step framework that um Ben and I, well Ben really developed and Ben and I’ve been using with organizations to sort of help them just really take all these ideas and all the sort of hype all the crazy and boil it down and get simple and get clear within. Inside your organization and take some concrete steps because there’s so much that is possible and so we really just wanna make sure that people are like how do I get out of like you know I’m in I’m paralyzed with the overwhelmed and get out of that space and really start taking some steps forward. All right, thank you. We’re gonna turn to uh Ben then since. We have a 5 step. We have a 5 step framework. We got a step by step roadmap, and then, uh, Kate will figure out, we’ll see how you, so it looks like 2 consultants and a and a client. Is that, is that essentially the, OK, OK. The client, the client is flanked. She’s a bit, um, she’s flanked. She can’t get out the center of the universe. Kate is sitting in the middle between Jen and Ben. Oh, Jen and Ben. Oh, I just, I just that it all works out. All right. Uh, alright, so Ben, why don’t you wanna acquaint us? Why don’t you just like take off the five steps and then. We have we have time to go into, but just tick off our five step framework for AI adoption, uh, that, that is a low hanging fruit, the low hanging fruit, that’s what we’re focused on. So we’re, you know, so many of our clients ask us how do we start, right? Because they’ve heard all about this AI stuff. How do you start? And so we’re trying to give people a real simple sort of process to follow, just the simple stuff. We’re not asking people to do big projects. We just want them to dive in to start to feel like they can get some workflow improvements, some efficiencies. So there’s really 5 things you gotta do. Gotta do them real quick. Let me see if I can remember them. If I can’t, Kate has done this in her real organization for in real life. So rather than it being all theoretical, she can give us the real down low, but that noise is Ben slamming his hand on the table for emphasis. We have to admonish Ben, stop, stop doing that. I like to add a lot of extra junk noise. I’m usually doing this that’s quiet, so yeah, she shakes her hands. I just hands by her head. That’s fine. That’s a lot of fuzz. OK, 5 steps ready. So I can do it. Number 1, learn. Spend 2 hours getting all your staff up to speed on the basic, uh, basics of what AI is, what it’s good for, what it’s not good for, OK. Number 2, jam sessions. Just brainstorm with people about their own jobs, not about AI, but about their own jobs. What do they hate doing? What’s annoying? What. Makes them groan when they think they have to do it, right? Step 3, toss every idea you’ve heard into a big list. So, step 3, big list. Step 4, prioritize, right? So, look at everything on that list, figure out what’s high effort, low impact versus high impact, low effort. Prioritize according to that. Step 5, pick your top 23 things and write up a 1-page description of it. Blueprint, what is it good for? Why are you doing it? What might the technology be involved? What, what technology might be involved in doing it, then you run with the top 23. That’s it. OK, all right, we’re gonna turn to Kate. What, what, what was uh new music USA facing before the, the innovation that uh that uh BFD go with it. I love it. I love it that BFC and uh and Firefly brought to you to the new music. Sure, so New Music USA, we are a national. Nonprofit organization that focuses primarily on grant making for musicians and arts organizations and we’re a team of 10 people do grant making to individuals, yes, individuals, ensembles, venues, and we also provide mentorship opportunities as well. So where we were, uh, I think some of our team was using tools like chat GPT on their own but without any guidelines or real understanding of how to use it. And our connection with BFC is they helped us build our website about five years ago and BFT BFT, yeah, as of now, as of now it’s over, but um I’ve been, I’ve been working with Ben and his team since I joined the organization in 2023 and uh Ben and I were just chatting about. AI in general and he offered a training for our organization back in the fall of 2024 so we all got together and just had a learning session with Ben about what large language models are and how they work and I think in the beginning we were all quite intimidated and we just didn’t know where to start. We know that these tools are important and that. Everybody is going to be using them and everyone is starting to use them now, but it’s, it’s hard to know where to start, especially if you are, you know, a team of 10 with lots of competing priorities and tasks. It can feel overwhelming so that’s, that’s where we started just, yeah, let’s start with, uh, with the, the first step in the framework. Let’s stick with you, the, the brainstorming. Uh, how did, everybody, everybody did everybody contribute? Was it so, so, uh, Ben and I worked together to basically decide which members of our team would make the most sense to, to brainstorm with. So we chose myself, director of communications, we chose my colleague Nathan, who’s our communications associate who manages our social media, and we chose our development manager Valerie, who does all of our institutional grant writing. And these were the 3 people we felt could. Identify tasks that are repetitive that require grunt work where simple AI tools could help just alleviate some of the the burden of repetitive tasks simple tasks very simple because we’re after low hanging fruit here yes we’re and we’re not trying to replace anybody’s jobs we’re trying to just cut back on repetitive tasks that take a lot of time. OK, uh, so Jen, um. Jam session number 2, step number 2 in the framework is, uh, I believe is that the jam sessions or that’s jam sessions. That’s right, jam sessions. So after brainstorming, then jam sessions. What, what’s, what’s the difference between a brainstorming session and the? I mean, that’s basically the same except for jam you’re eating, obviously. No, I’m just kidding. Um, did I, did I mix up the framework? Step one is step one’s learn, learn. Oh learn. Oh, see, well, you’re all suffering with a lackluster host and Kate, Kate, Kate was too polite to say. You’re wrong. Uh, step, step one is not brainstorm. All right. Step one is learn. So step one, alright, so, OK, they did the learning. She said they, they talked to Ben. They got some training. That was the biggest thing like what is this and what does it do and what does it not do? Once you have that framework, then you can have your jam session because then you’re like, oh, OK, these ideas are percolating. But to Ben’s point. You’re not immediately diving into OK how is all that stuff I just learned gonna help me you actually stop and come back and go, what’s going on in my job? What do I hate doing? What are the things that I take like this this sort of grunt work, the repetitive task what’s stuff that I’m like oh my god, I hate sitting down with a blank slate looking at. Um, emails I need to go right back to all my grantees. I know they’re gonna be, it’s the same stuff every time, but it takes me a while to find everybody, do the things, write the note to everybody, whatever. I have to go research. I have to go dig through papers. I do whatever that’s like a lot of just sort of like, uh, I hate that part of my job. So you don’t even really get into like, can AII help me with this yet? Like Ben said, you’re just really like, let’s talk to everybody about their jobs. So the jam sessions are more about like, hey, let’s just actually talk about your day to day and what’s the stuff that’s sort of is a drag. That’s your jam session. I know it doesn’t sound like a jam session, but it is because it’s, it’s also very cathartic because you get to like talk it out with your, with your teammates and be like, Oh, this is kind of a drag, and you’re like, Oh dang, OK, great, I can help with that. But let’s not get to that yet. Let’s just say, tell me what you actually literally don’t love about your job. Just get it out. It’s good therapy. It’s get it out, and then you can be like, Great, I could probably help you with that. OK. And, and it’s new Music USA, so a jam session is, is appropriate. It could be even every day. Alright, um, OK, so now we have our big list. We’ll go back to Ben, back to Ben on the end. Um, what, what’s, what’s our big list, uh, step look like? So big list, there’s the normal way of doing this and then there’s the technologically you get like an A plus way if you do it, which is to use AI to help you. With you, OK, we don’t even wanna go there necessarily normal. The normal, the normal way is to listen to, to people in jam sessions and any time an idea comes up, hey, I wish, I, I hate, like Jen, you know, like Jen said, I hate writing emails to all my grantees because they’re all the same. There’s there’s a little bit different thing in each one, which is kind of annoying. You write that down, that’s all. Write it down and make a list and so as you do that by the time you’ve done these jam sessions you’ve hopefully got a list of, you know, 1520 things on it. So that’s the base, that’s the normal way, OK? If you want an A plus in AI stuff, OK, OK, the AI version would be to go to, you know, a quad or something like that. Record all the meetings, by the way, that’s the key. So record the jam sessions, then you download the transcript, right? And you’ve got all the words in the transcript, and that’s all these AI tools work on. They just work on words. So you’ve got all the words in the transcript. You take the transcript, dump, dump it into cloud or JJBT, whichever one you wanna use, doesn’t matter which one, and you say. Hey, I’m doing a jam session, and the purpose of this jam session is to find everyone’s grunt work, right? The things they hate doing, the repetitive, the road stuff that maybe AI might help with, and can you read through these transcripts and pick out ideas that you find? There you go. So you could do it the computer and the AA plus plus plus way is to do it both ways and then combine them and compare, yeah, compare and combine. OK. Do we know yet whether all the items on our On our, on our, uh, big list are are AI achievable? We do not know. We do not know, and a lot of times they won’t be exactly. So that’s what the next, so then you go on to the next step, right, the prioritized step. So I don’t want to skip ahead on my steps. OK. Where, where is there anything more you want to say about, uh, about step three? The, the, the big list? No, I don’t think so. I mean, we covered the big list, list, how, how, how many items do you remember? What, what was your, so, or between, so, so Ben met individually with me and my two colleagues, and between the three of us we came up with a list of 15 possible use cases. Give us a couple of examples of. Sure, uh, well, the easiest one is an AI note taker, um, but things like, so we publicize our grantees on our website to do that we have to create a lot of profiles for each grantee, so like 150 profiles per year in WordPress, so very simple repetitive task that takes a lot of time, that’s one, an AI writing partner for social media strategy is another example. Let’s see what else one more if you can think of one, sure, uh, a knowledge base for grant metrics. So basically loading all of our existing grants into cloud and then making it easier for our grant writer to find metrics from past grants and update them just to save a lot of time. OK, all right, cool, uh, now we’re prioritizing Jen. Why, why don’t you, uh, lead us through this, I think that, you know, honestly that’s a big one where you, I mean, again, the list that you’re making usually isn’t like, well I’m gonna do this in cloude or I’m gonna do this in whatever you just say like I really hate it when you’re, you’re. She’s already gone through all the steps, but when you’re first doing it, you’re usually just saying, God, I really, I need to, I need to go back through all of our past grant applications or I wish I could go back through all my past, you know, applications and pull really great information out of that. God, I wish there was a way to do that. Then you figure out that AI could do that so like we said, there’s usually a list of things you have to go through them and figure out which ones are actually going to be kind of the most AI. Friendly, you know, like these are gonna be the achievable, but back to Ben’s point, you’re literally looking at like high impact, low effort, so you have your think of a little grid. You’re like impact over here and effort over here and one is X and one is Y, right? So you’re looking at high end so you then as a group you kind of map them out and you go how many people in the org will be impacted if we implement this solution that’s gonna be a high impact solution and it’s actually not gonna be that hard obviously rolling out. AI note taker is like the lowest effort thing you could possibly do, and it will have a huge impact on everybody in the entire organization if you roll it out properly so that’s the easy like boom boom it’s it’s a it’s a double win, you know if you go, I wanna create some fairly complicated custom, you know, thing or whatever you’re like oh that’s a high effort and it’s only really gonna impact maybe one or two people on the team. We’re gonna put that in the let’s get to that later list so that’s where you, you make a grid and you put you plot all 15 things out on that grid with some maybe some help from then your consultants because that you’re like I don’t know how difficult this is to implement and so that’s where really the. The folks who’ve been doing this work can kind of jump in and be like, oh, that’s actually a pretty high effort. It’s not, uh, I know all the, all the marketing says it’s really easy, but actually that’s gonna take a little bit more time than some of these others. So helping a group sort of sift those things onto the grid, it’s easy to sort of pick out the top 23 if you’re ambitious, maybe 4, of those high impact, low effort. How long is this 5 step, uh, framework take to work through at an organization that has like New Music USA 10, you said 1010 employees. OK, so how long did it take you? Well, it’s still very much in progress. I mean, you know, like, like everything, it’s a matter of, uh, prioritization so I mean. You start then, when did we start? Was it we really started in earnest in the fall, right? Yeah, so it’s been 8 months or so, yeah, roughly 6 months, yeah, I would say about 6, yeah, yeah, maybe even less. And now you’re at the point where you’ve got your list, yes, so we are now starting the implementation phase. We, we have our, our top 4 that I alluded to, and we’re gonna start with implementing 2 of those 4. OK, so you’re at the top 2 to 3, uh, and the blueprint. You’re in the blueprint phase. What’s, what’s this blueprint all about, uh, Ben? Well, the blueprint blueprint is supposed to be more detail on each idea, and you want to capture, uh, what you’re what you’re trying to replace. So what is the pain point that you’re trying to replace? So what is the thing that people hate doing? That AI is gonna do, um, how’s it gonna do it for you and hopefully a little bit about how it’s gonna be implemented again, you might need some contractor help or, you know, do some research on this, but a lot of times it’s gonna be, hey, we just have a custom prompt that we’re gonna put into a cloud. Other times it might be, hey, we need to connect Zapier with in a. AI system or something like that. Um, but you just want a page or two on each idea and the reason you want that is so that you can take it and hand it to somebody to implement for you. So it’s either maybe somebody on your team or in your organization or it’s somebody you need to hire from outside, um, but it’s like at least a sort of a concrete, you know, not just what this why this exists and what it’s gonna do. But a little bit about how, how it should be done as well and if I may add, I think the blueprint is also very helpful if you’re in a position where you need to get leadership buy-in for implementing such a tool. Nonprofits are slow to adopt to new technologies which is, uh, and I mean arts nonprofits I think are even slower in many cases. So, um, you know, having this blueprint was extremely helpful. I was able to send it to our CEO and say. These are 4 tasks that we can undertake. This is how much time it’s going to take to implement them and how much time it could save us in the long run. yeah, yeah, exactly, and, uh, metrics like this are so important and the blueprint really helped establish those and helped me get leadership buy-in so and then it’s uh implementation. OK, OK, um, you’re gonna talk some about ethical issues around AI, I believe that’s what, that’s what you’re, uh, responsible adoption, there’s a little bit of responsible adoption, yeah, flesh that out for us. I mean, essentially we, there’s, I think we kind of because we only have a 30 minute session we can’t touch on everything, but we do talk about, um, at the beginning a little bit of just about again like backing up to saying like we understand. There are a lot of considerations to go into using these technologies um there’s a lot of um focus right now on the environmental impacts of data centers and things like that so um and. The politics or the. Ideologies of the folks who run the various uh companies that do um the large language models that we’re using right now, the generative AI that most folks are using, so mostly we just wanna bring um some of those ideas, uh, to the fore. We don’t have a lot of time in our session to talk about it, but we can do a whole other, I mean, I’m sure there are other sessions at the NTC this year about the ethics, um, around. Using the tools so it’s also there’s a lot of thoughts about like uh what does it still mean to be the original creator of something? What does it mean now when all the ideas of all of humanity are basically in a giant robot machine and it’s you can spit it back out of you so there’s all these different sort of ethical theoretical ideological considerations, environmental, all these things and it’s just important for us to help people understand. In the learning part we very much say go to some learnings that actually really dig into some of these things if you don’t know about these things already you should educate yourself and then. Like in a session I had on my untangled a couple weeks ago, it was like talk about these things as a group inside your organization. She drops the name I did the podcast. I did, but she did it so obliquely that it’s, probably not even noticed. I just called it out, of course, but man, it’s called, it’s called Untangled. Brian Miller and I were talking about exactly these things to do before you get going, which is to say as an. Organization, where are we with our values and what kind of, um, you know, organizations do we want to work with whether it’s the AI tool or the vendor that you know prints our posters or whatever we need to think about that as an organization you probably already have those conversations about other vendors or other folks you wanna partner with your AI tools are no different, so it’s like how do we want to take our ethics and our values and make a values aligned decision about what tools. We’re going to use and how we’re going to use them as an organization. I like your analogy about working with other partners, consultants, partnering with other nonprofits. I mean you would, you would consider their, their values. You would, uh, you should do the same here. This is, this is a partner. That’s, that’s, that’s a great analogy. And let me, let me add one thing about our framework. We’re about to wrap up now. I want to thank Jen Frazier, Ken, no. I got him. He’s got him. He got him. He’s like, wait, what? I was like, Oh, my mic’s off. I think I turned it off. It happens every time. Please, please, um, so framework wise, by the way, you can actually download our framework from our session notes from our session if you want. We have a little framework handout. Ethics is captured in part 5, OK, which is the blueprint and execution phase. As you do that, you want to, every time you implement a new tool, you wanna update. This thing which you should have called an AI use policy for your organization and so the process of creating one of those should hopefully capture your ethical stance on this stuff you should consider what you wanna use, what you don’t wanna use and put that into your AI use case or AI use policy. Every time you finish, you hit, you know, step 5 and you start executing on your ideas, you wanna go back to the AIU’s policy and add whatever tool that you’ve created or adjust what you have in that policy to reflect the tool that you have adopted or the thing that you’ve created. Can we spend a couple of minutes on the environmental impact? I mean. The environmental impact is unavoidable as soon as you start entering prompts and then massive data centers are churning through your prompt and going through their quadrillions of bits of data and so the the. The environmental impact is unavoidable from, from, from the get-go. If, if you are uh aligned with. The, uh, you know, sort of climate change mindset and that climate change is a is a is an issue for your, for your nonprofit, even if it’s not your work, but it’s part of your values, then, then you have a, you have an unavoidable. How do you reconcile that conundrum that you do that’s uh, you know, a dissonance. Yeah, so, so this is how I think about it. Everybody should think about it in their own way, do their own research, figure out, which is a terrible phrase to use, but do research, figure out how you feel about it. So. The way I think about it is there are environmental trade-offs about everything we do, so drive a car, right? You should probably if you can take the bus, take public transit, not contribute to climate change, right? But there are good things about cars. There are ambulances, for instance, which can take people to a hospital, so it depends on. What you’re using it for, right? So do you wanna use AI to write yourself jokes to make you laugh before bed? Probably not. Are you using AI to change the world in a better way, uh, and that’s gonna be super impactful then you know you wanna think about it. The other thing to think about is the context of the environmental, the actual, you know, the actual data about what. How, how the, the data centers impact the environment is a bit opaque. That’s I think the biggest problem with using these tools is that we actually don’t really know how much energy they use and how much water they use. It’s hard to extract that information from the, you know, the systems that are in place. On the other hand, we do know that Google’s largest data center last year. Used as much water as the average golf course in the United States. That’s the golf course. So what’s more useful? What’s more you know, what contributes more to the world? I would argue, you know, a data center, but maybe not, you know, it depends on your perspective. If you’re a golfer, maybe not, but it’s, I think about all the golf courses on the planet and the. Places also where golf courses are. Let’s go with Palm Springs or Arizona. Same thing. It’s like ridiculous amounts of water that goes into a very luxurious, I think, um, elite elitist, yeah, elitist sport that’s very expensive, blah blah blah. So there’s that, there’s, um, the environmental impacts generally right now are, um. Water consumption is the biggest thing and obviously um but there’s other impacts about like where are they building the data centers who are who’s actually staffing the data centers what people are what folks what’s happening in that community a lot of folks are actually fighting. The building of data centers in their local communities and things like that. So there’s, you have to kind of like dig into it. But to Ben’s point, it’s very hard to find this information. They like to hide it from you as much as they can, but there are different ways. So I’ll just drop another name. There’s an organization here called Change Agent, which is an open source, uh, based and very, uh, uh, an AI tool that’s just like, uh, Chach BT or Cloud, but. They’re built from and within the social sector so shout out to those guys they will give you a rundown of how they are combating the environmental impacts of the data centers that they’re using to process their stuff and so they do a lot of carbon offsets they they do a lot of work internally as an organization to already think about this and combat that so they’re again because they’re from the space they’re already thinking about this and they’re like hey. We know this is a problem we have ethical problems with it, but yet we also see huge benefits from organizations being able to use these tools, so they’re a good one to also just look at if you’re considering AI tools. And Kate, I’m gonna give you the closing word, but how did the New Music USA reconcile the the environmental impact with the, the return on investment that you that it expects to enjoy from. I think using the uh yeah picking up the low hanging fruit with AI sure I mean I think the way we’re thinking about it is more in terms of you know these tools are here and are going to be broadly used and adopted by most organizations and most people so our take on it has been how do we learn to use these tools that are here as responsibly as we can use them. All right, perfect way to start. That’s, uh, to end. That’s uh Kate Dreyfus, director of communications at New Music USA. With Kate is Jen Frazier, CEO and founder at Firefly Partners, and Ben Freda, president of BFC Open Peren BFD Close Peren Digital, whichever you prefer. Now, the name of the company is BFC Digital. Ben Freda. All right, Jen, Kate, Ben, thank you very much. Thank you and thank you for being with Tony Martignetti nonprofit radio coverage of the 2026 nonprofit Technology conference. It’s time for Tony’s take 2. Thank you, Kate. We have a new character in our ensemble at the gym. Community gym that I go to. Uh, this is Sam. Now, Sam, uh, is the gentleman who invited me to sign Jerry’s birthday card. Remember Jerry, who I don’t know, but I recognized him when Sam told me where he sits because he sits at the same bike every single time and he’s turned 91. I signed his card. That was Sam who invited me to do that, not because Sam knows my name, but just because he sees me a lot. So, he invited me to sign. Sam is, uh, Sam is interesting. He’s very. Uh, talkative. Um, he’s, he’s, I think he’s north of 80. He’s got a, this great shock of white hair, all pure white hair, slender, tall. I believe he was in the military. I think I’ve heard him say through the years that I’ve been going, I, I, I, he was in the military. You may have even flown, like a navy, maybe a navy pilot or something, a navy flyer, not necessarily pilot, it could be backseat, some aircraft, but Um, I have to, I have to get a little more bearing on, on Sam. So I just, but I just learned his name. That’s why I’m introducing you to him. I didn’t know who he was, but I heard, uh, him introduce himself to someone else. Yeah, he’s, he’s, he’s chatty. Um, like, he walks around to the different stations and different machines and people and, I don’t know if he’s interrupting, but he’s, he’s imposing. You know, he wants to talk about, he talks about sports, he talks about politics. Uh, he talks about, uh, town stuff, like, you know, events that are coming up, you know, like we had, uh, we had, uh, Saint Patrick’s Day last month, festival, things like that. But he’s, he goes around and he’s, uh, he, he, he is slender, but it’s not because he works out a lot because he’s, he does more walking in the, in the fitness center than he does actually working out at machines or weights or anything. Uh, he’s, um, He, uh, he, he likes to, he likes to tell stories and talk to people. So. That’s Chatty Sam, very nice guy. Well, if you can put aside the food aside the fact that he’s imposing on everyone’s workouts. He doesn’t impose on mine. Uh, and because I’m on the elliptical, or I’m down on the floor and I’m panting as I’m doing my planks and so, plus you’d have to bend down to talk to me. So I’m inconvenient. But, uh, he talks to a lot of people, unless you’re, I noticed, unless you’re on the treadmill, the treadmill, he, uh, he avoids. But if you’re at any of the machines, or even on the bikes, He comes over for a chat. It’s a little much, a little much, uh, I believe. That’s Chatty Sam. And that’s Tony’s take too. Kate So do we know if the guy ever got his birthday card? Oh, Jerry did get his birthday card. Yes, but it was about. 2 weeks late, because he didn’t come in at the same time Sam was coming in with the card. So by the time they got together, it was about 2 weeks after. Uh, Jerry’s birthday. It’s still the thought that counts. Absolutely, absolutely. And he’s probably still, I’m sure Jerry’s still scratching his head. It keeps him up at night. Who’s the person who signed Jim Rat? Who could that be? Who’s Jim Rat? We’ve got just about a butt load more time. Here is ethically using AI. Welcome back to Tony Martignetti nonprofit radio coverage of 26 NTC. That’s the 2026 nonprofit Technology conference. We are all gathered, all these technologists and technology users in, uh, Detroit, Michigan. With me now are Chris Rossica, CEO and president at Rossica Communications, and Bethany Friedlander, president and CEO at New Bridge Cleveland. Chris, Bethany, welcome. Thanks, Tony. Thank you. Pleasure. Have you both. Your session topic is 5 tips for ethically using AI to develop compelling website and social content. Chris, I’m gonna ask you to just give us a high level overview to start us off. Sure. Our room was pretty packed with mainly nonprofit executives who were really based on the title of our topic looking for ideas and how they could use AI to generate content uh whether it’s for their for an article or a blog, how to repurpose that content for social media. Also, how to leverage earned media coverage, say, um. An interview with Bethany Friedlander from Newbridge and take that interview and turn it into say an e-newsletter for stakeholders so it was really about utilizing AI in ethical ways of course but utilizing it to really integrate. And diversify marketing communications so it’s less taxing on staff to do that work and it makes it more feasible to do the work that in the past may not have been a priority based on. Resources resource limitations and lack of qualified personnel to, to do that work. All right, thank you. That’s ideal. Thank you for that, um. Now, uh, Bethany, are you a, a client of Rossica Communications? Is that the relationship here, client and consultant? Yes, we actually met at South by Southwest EDU. We sat next to each other at a breakfast and have developed a really strong relationship, um, primarily working on elevating the story of New Bridge Cleveland. We’re a smaller nonprofit. Uh, and so, uh, looking to frame that story, how do you become not a nice to fund but a must fund, and how do you do that, particularly bringing national money into Cleveland, which is not a place where a lot of national money finds its way actually we have a very robust local philanthropy but not nationally. OK, and, uh, before working with, uh, with Chris and Rossica, what, what, what was your experience, if, if any, with artificial intelligence using. I think we were probably in the same place a lot of nonprofits are, which was, is this cheating, right? We’ve, we answered the same grants 15 years in a row with the same questions. They’ve never changed the application. Um, is it cheating to use AI? And I would say no, we have our own blinders on about the language we use to discuss ourselves, and I think it’s important to have an opportunity to have that looked at, challenged, and improved. OK, um, Chris, why don’t you, uh, walk us through, you have. You have 5 strategies of uh of uh ethically using AI for web and and social content. Can you talk can you just tick those off and then we’ll talk through? Sure, sure. So the ways that you can use AI include positioning and messaging. And help frame to frame the story. Uh, the second is for repurposing. Content for stakeholder communications. Uh, the third way or that you can utilize AI is to help you expand your website content, and that includes articles, thought leadership articles on your site. The 4th is to utilize these tools to help you. Be more searchable online. So AI search today, Tony is huge and. Many nonprofits were struggling, are struggling or still struggling with search engine optimization. Now you have AI search and how do you rank, how do you rank in AI search? So that was another key topic we spoke on and then the last is really creating your brand persona and how do you stand apart? How do you differentiate your organization? OK, OK. Um, around the, uh, the, the ideation, you know, thinking about the brainstorming, how you might have started using AI at, at, at New Bridge Cleveland, Bethany, what did you, was there a brainstorming session or, or meetings or like how did you decide where, where the, the. The best use cases would, would be to begin. Uh, so a number. So first of all. So, a lot of different opportunities, right? So one of the biggest for me was how do you choose character limitations and grants. So you can write the most eloquent answer in the world and if it doesn’t fit within 250 characters, so it’s fantastic as a tool to take you and edit as an editor. Also, to, um, help you brainstorm, so the opposite, right? So sometimes I know exactly what I wanna say and I’m not saying it in a few enough characters and sometimes I don’t know what I wanna say at all and so I can start in a very casual way throwing ideas and and I and words into into AI and then getting back the brainstorming that then I can then further refine. Chris, any, any thoughts on this ideation phase? I think that. Having AI really get to know you and what you’re all about is critical and Bethany’s done a great job of doing that, of educating the AI of prompting it, of having it really explore what her brand is all about. Most websites today aren’t up to date with people’s messaging. Why? Because all organizations evolve. All nonprofits are, are evolving. The landscape is very it’s challenging right now because federal funding has been cut across the board. And how do you stand apart and diversify your funding mix, right? And that’s one of the things that New Bridge has done a great job with it’s. Really giving AI all of the information and they’ve also updated the website too recently with very current messaging um again many fall behind in that area, not just nonprofits but across the board corporations too as well so that’s something that I’ve seen really pay dividends to Newbridge and Bethany I’m sure has a lot. To say about how she works to inform the AI and really let it get to know her. So I throw every email that I write into it with the quote uh with the prompt of make this better. I may or may not use the change, but every time I do that it’s learning my language. It’s learning my phraseology so that then when I ask it to create something new, it’s creating it more and more frequently and tighter and tighter closer to my own voice. We’re also a fairly complicated nonprofit in terms of how we do the work, and that’s been a challenge since our inception to explain to funders and so. Allowing, uh, myself to engage in a dialogue with AI to say which parts of this are the most resonant with the uh with the with the audience that I’m trying to get to and not. And understanding that it really only matters that it’s heard, it doesn’t matter what I’m necessarily prioritizing, so letting it do some of that work for me and acknowledging that I may not be the best lens, we’re very close to our work. I don’t think there’s a nonprofit here who isn’t deeply passionate about their mission, deeply resonant. That can be a huge limitation to your success because it puts huge blinders on you and the language you use to describe your work. That’s interesting, yeah, yeah, um, when you’re, uh, when you’re uploading everything, including your emails, are, are you, you, you’re doing this to a, to a learning base that is unique to, uh, to Newbridge? It’s not, this is not in the public domain, right? Well, and that’s a funny question because I am giving it a tremendous amount of power and information that it can use then to improve other nonprofits, right? But I think it’s still worth it to have that refinement. To, um, and I would say now, uh, I would say about 85% of what it writes. I clearly hear my own voice. I, I no longer question it does not seem like a foreign entity, but that’s also because I have my own account. I don’t let anyone else use my account, so it’s only getting input from me. It’s not getting input from the entire organization. OK, OK, but, but it is one of the public models. Yes, it is, it is. OK, and so that doesn’t, I mean, every email, maybe not every email literally, but you know if it’s 95% of your emails or even 75% of your emails, that that doesn’t bother you that. That they’re now in the, is it right to say they’re in the public domain? Well, they’re they’re they’re in the domain that that this large language model which what what do you use? chat, OK, that chat is learning from that doesn’t, I mean that that doesn’t trouble you. It does. I suppose suppose somebody does a prompt write me an email in Debbie Friedlander’s tone to. To uh have her director of finance write a check for $50,000 to, to my company. I mean, yeah, no, I think it’s a fair question, and we had a great audience member who wouldn’t talk about like the ethics of all of this, and I think it’s incredibly important. I think where my head’s at right now is a 15 person staff. With a $2.5 million dollar budget that has to be raised like clockwork starting July 1st of every year and right now the benefits outweigh the liabilities, I care deeply about my participants and making sure that no private information about my participants is ever included in those emails. I would never do that. Uh, I would also never put an email in that had any private information about a funder or an arrangement with a funder, but if we’re talking about, um, you know, conversational emails where I’m talking about Newbridge and. Every day I have 10 opportunities to describe the organization. They don’t all land equally and so being able to run it through that and recognizing. I think it’s elevated the language and I mean my proof is, is that it’s elevated our status in Cleveland. It’s it’s increased the number of tables that we’re sitting at. It’s increasing the number of dialogues we’re participating in, and I think ultimately that pays off. That’s interesting. So you, you, you’re seeing those kinds of outcomes, prominence in the city, absolutely. Being a thought leader and I mean you can only be a thought leader if your thoughts are well organized and I don’t know that that’s something that naturally we all do particularly well and heard and heard and heard otherwise you you if your thoughts are well organized and nobody reads them, that’s called a diary you know. It’s not what you’re aspiring to, right, uh, OK, no, but you’re right, there’s risk and there’s, and you are enriching something that you’re never gonna get back, so you have to hope that your returns coming in another way, and I think right now benefits outweigh liabilities. I, I can’t say that that’s gonna be true forever, OK. Uh, Chris, say something about the, the AEO I call AEO, artificial engine optimization. You, you said earlier a lot of nonprofits are still working on SEO, but that, that really has been, uh, overtaken. Uh, it’s almost an anachronism now because of the, because the artificial engine, the AI engines are generating summaries that don’t, don’t lead to your website. How do we overcome this? Well, certainly their new environment. There are a lot of It’s pretty simple and straightforward ways of understanding how AI works. And it’s not really reverse engineering either. You can look at who is ranking for certain key phrases so if you put in workforce development, healthcare workforce development organization in Cleveland. We better come up first. You better come up first. Exactly. This relationship is not working otherwise. Exactly. You’ve been admonished. Yeah. So if You put in a cer a certain key phrase depending on where you’re located and. The service. The geographic surface area that you’re serving, right? You can be a national organization most of the time though, national organizations have a local presence, right? So the very the telltale signs are, you know, are you ranking in AI search results and if not. Look, scroll down and then you’ll see people who are searching AI also ask and then you’ll see some additional key points. So you’ll know the important, that you’ll know the ways people are searching in AI for answers. After you understand that, then it’s time to modify your website and the content on your website to create. Summaries, really succinct summaries and Q&A’s or FAQs that address those questions that are being asked on a regular basis. So that’s, these are just some very basic, uh, ways of beginning to appear in AI search results. There are some other things that are a little bit more technical. So if any of your listeners have questions about that, we’re very transparent and very willing to share with them those specifics so if they have a web developer that’s part time or an agent, a company that’s working for them, we can steer them in the right direction and let them know what they need to do so they rank in AI search responses. OK, OK, um. Bethany, how about, uh, expanded web content? You, I guess this is a, this is generating website content for you as well as social, right, as well as social content. Yeah, so before we met, Chris, we had no social media presence. I just didn’t have anybody. I feel, well, let me take a step back. I think it’s really important if you’re gonna add anything to your organization you need. To be doing it consistently, so having, you know, a Twitter account if you’re tweeting tweeting or Xingxing I guess Xing once a month is is irrelevant so I never had the staff to do it. Chris has showed us how we can take content that we put in one place and use it in multiple other places in a really simple way and so we’ve greatly expanded our footprint and I think um it’s really about. It’s not, it’s not about eliminating work, it’s about doing work smarter. And so I think we were in the phase of just we can’t, and now it’s we can and we can do that smarter. And is that AI related? It is. So explain how you took, you can take like one of my articles and then so Bethany was featured on Medium or Authority Magazine’s Medium platform. Millions of people have exposure to that, but how do you know if the right people are. You know, tuning into the website and seeing Bethany’s uh bylined article or op ed piece. So what we did is took use utilizing AI took that interview with Authority magazine, turned it into a blog post that went on to New Bridge’s website, turned it into a LinkedIn social post, uh, all in an automated fashion that saved so much time, and there’s no. It’s in the, the information. The article was in the public domain to begin with. It’s on Medium.com. So I understand the question you asked earlier, Tony, about sensitivity and do you really want to give it confidential information and 99%, I mean, almost all of what we do is not that at all. We’re, we’re, we’re very wary of doing that and I concur. I get the sense that you feel you, you have some concerns about that. And we have the same concerns. I know Newbridge does too, but in this area. That article. Now it’s turned into all this content for different channels and it’s all unique content. Same holds true not only for the earned media coverage, but for an article, a thought leadership article that is written for the Newbridge website, that can be turned into a media pitch. Well that can be turned into a um into social posts for all your social channels and then we use AI on an ongoing basis to take one social post that we write and turn it into social posts for all of uh nonprofits channels um so it’s unique content but it’s still conveying the same message. You, you had your session already, right? Yeah. So what are some of the questions you got that struck you that are memorable? So we, we asked them a question which was to put, uh, a simple question into whatever uh AI they’re using, which is what movie star is my organization and. Um, and I do that and I talk about that because it allows you to see how you’re being seen. It’s the language you’re using. So interestingly enough, overwhelmingly in the room, everybody got Tom Hanks, and the, and the inclination was, oh, then AI is not that smart, it only has one actor, and my response is no, that’s because we’re all talking about our organizations in the same way. We’re all talking about ourselves as kind, good people. Now, interestingly enough, Newbridge doesn’t come up that way. Newbridge comes up as Viola Davis, and that’s partly because we talk a lot about trauma, resiliency, and grit. And if you wanna talk about somebody who embodies that, Viola Davis really embodies that. So my suggestion to people is if you don’t like what you’re getting back, then think about how you’re talking about yourself because we’re really much more complex than just being good. By the way, Rossica Communications is George Clooney, so which, which was a shocker, by the way. Clooney. OK, now I’m thinking of Alec Baldwin. Yeah, I was thinking of the personal side of Alec Baldwin. No, George Clooney. Now that, yeah, so are you happy with that? or very happy, yeah, OK, yeah, great under pressure. He always seems to be great under pressure. His persona, right? His acting persona, very consistent. You can rely on him, yeah, those types of attributes, yeah, does similar kinds of roles too. Yes, right, OK, leading man, very good. OK, uh, what’s, what’s another question you got that might be interesting folks? So we also ask, uh, what’s something about my organization that’s confusing or might be misunderstood by funders. Again, something that you may not see, it may be a huge blind spot for you, but can help you with clarity and then also what is the number one thing we could be renowned for? What can we be world best at? So I think. Um, asking it those questions, it’s not, I mean, it, it’s not foolproof, but it does give you a sense of how somebody who’s not with you every day might respond to you, think about you, and what they might be left with is an impression. If it’s not the impression you want them to be left with, change the language you’re using. Yeah, Tony, we always say perception is reality. And how are you effectively. Demonstrating the value you bring to the stakeholders that you serve and if you’re not differentiating your organization and if you’re not. Saying things that truly resonate with those stakeholder groups, then you’re missing a huge opportunity to, to improve communications in order to serve more people that you’re looking to serve. How about we leave it there. All right, thanks. That’s Chris Rossico, CEO and president at Rossico Communications, and Bethany Friedlander, president and CEO at New Bridge Cleveland. Chris, Bethany, thank you very much. Thank you. Thanks for being with us and thank you for being with Tony Martignetti nonprofit radio coverage of the 2026 nonprofit Technology conference. Next week, we’ve got more on AI from 26 NTC with AI for the rest of us, and your AI acceptable use policy. If you missed any part of this week’s show, I beseech you, find it at Tony Martignetti.com. Our creative producer is Claire Meyerhoff. I’m your associate producer Kate Martinetti. The show social media is by Susan Chavez. Mark Silverman is our web guy, and this music is by Scott Stein. Thank you for that affirmation, Scotty. Be with us next week for nonprofit radio. Big nonprofit ideas for the other 95%. Go out and be great. OK, um, a little note here. We, when you, we’ve got more, the way you said more made it sound like it’s, it’s too much. It was, it lost the effect that we’ve got, like, we’ve got more. All right, you didn’t, I’m I’m exaggerating. You didn’t do it that much, but just, we’ve got more on AI. Please, just do next week for me. Next week, we’ve got more on AI from 26 NTC with AI for the rest of us, and your AI acceptable use policy.

Nonprofit Radio for May 26, 2025: Healthier Productivity From AI

 

Jason Shim & Meico Marquette WhitlockHealthier Productivity From AI

Our annual duo returns with tips and resources to make your use of artificial intelligence better for you. They also go beyond AI with many smartphone strategies, inbox management, and Meico shares his shutdown ritual for bedtime. They’re Jason Shim, from Canadian Centre for Nonprofit Digital Resilience, and Meico Marquette Whitlock, The Mindful Techie. This is part of our coverage of the 2025 Nonprofit Technology Conference (#25NTC).

 

 

 

 

 

Listen to the podcast

Get Nonprofit Radio insider alerts

 

Apple Podcast button

 

 

 

We’re the #1 Podcast for Nonprofits, With 13,000+ Weekly Listeners

Board relations. Fundraising. Volunteer management. Prospect research. Legal compliance. Accounting. Finance. Investments. Donor relations. Public relations. Marketing. Technology. Social media.

Every nonprofit struggles with these issues. Big nonprofits hire experts. The other 95% listen to Tony Martignetti Nonprofit Radio. Trusted experts and leading thinkers join me each week to tackle the tough issues. If you have big dreams but a small budget, you have a home at Tony Martignetti Nonprofit Radio.

Nonprofit Radio for May 5, 2025: PII In The Age Of AI & Balance AI Ethics And Innovation

Kim Snyder & Shauna Dillavou: PII In The Age Of AI

Artificial Intelligence and big data have transformed privacy risks by enabling malicious, targeted communications to your team that seem authentic because they contain highly accurate information. Kim Snyder and Shauna Dillavou explain the risks your nonprofit faces and what you can do to protect your mission. Kim is from RoundTable Technology and Shauna is CEO of Brightlines. This continues our coverage of the 2025 Nonprofit Technology Conference (#25NTC).

 

Gozi EgbuonuBalance AI Ethics And Innovation

Gozi Egbuonu encourages you to adopt Artificial Intelligence responsibly, in a human-centered approach. First, be thoughtful with the threshold question, “Should we use AI?” If you go ahead: Create a thorough use policy; overcome common challenges like staff training and identifying champions; manage change intentionally; and more. Gozi is with Technology Association of Grantmakers. This is also part of our #25NTC coverage.

 

Listen to the podcast

Get Nonprofit Radio insider alerts

Apple Podcast button

 

 

 

We’re the #1 Podcast for Nonprofits, With 13,000+ Weekly Listeners

Board relations. Fundraising. Volunteer management. Prospect research. Legal compliance. Accounting. Finance. Investments. Donor relations. Public relations. Marketing. Technology. Social media.

Every nonprofit struggles with these issues. Big nonprofits hire experts. The other 95% listen to Tony Martignetti Nonprofit Radio. Trusted experts and leading thinkers join me each week to tackle the tough issues. If you have big dreams but a small budget, you have a home at Tony Martignetti Nonprofit Radio.
View Full Transcript

Welcome to Tony Martignetti Nonprofit Radio, big nonprofit ideas for the other 95%. I’m your aptly named host and the podfather of your favorite hebdominal podcast. Oh, I’m glad you’re with us. I’d turned dromatropic if you unnerved me with the idea that you missed this week’s show. Here’s our associate producer Kate to introduce it. Hey, Tony. Our 25 NTC coverage continues with. PII in the age of AI. Artificial intelligence and big data have transformed privacy risks by enabling malicious targeted communications to your team that seem authentic because they contain highly accurate information. Kim Snyder and Shawna Deleu explain the risks your nonprofit faces and what you can do to protect your mission. Kim is from Round Table Technology, and Shawna is CEO of Bright Lines. Then Balance AI ethics and innovation. Gozi Egbuonu encourages you to adopt artificial intelligence responsibly in a human-centered approach. First, be thoughtful with the threshold question. Should we use AI? If you go ahead, create a thorough use policy, overcome common challenges like staff training and identifying champions, manage change intentionally, and more. Gozi is with Technology Association of Grantmakers. On Tony’s take 2. Tales from the gym in addition to my gratitudes. Here is PII in the age of AI. Hello and welcome to Tony Martignetti Nonprofit Radio coverage of 25 NTC, the nonprofit Technology Conference. We’re all together at the Baltimore Convention Center, where our coverage of 25 NTC is sponsored by Heller Consulting Technology services for nonprofits. Our subject right now is PII in the age of AI. Personally identifiable information in the age of artificial intelligence, safeguarding privacy in a data powered world plus we’re adding in the topic. Alright, already the show’s over. I wanna thank you all for coming. Uh, we’re, we’re here all week. Uh, be sure to tip your servers, um, and we’re adding in the topic a little more privacy please. Colin, diving into data privacy. All right, because, uh, our guests, um. Ask to combine topics which made a lot of sense. Um, but, uh, before I introduce the guest, well, now, let’s do it this way. So we have, uh, stand by there. We have, uh, first is, uh, Kim Snyder. Kim Snyder, um. I gotta take a deep breath. I do, uh, Kim’s title. I’m gonna hyperventilate trying to get enough air to oxygen in. I’m only 140 pounds. I don’t carry enough in my lungs to carry this, to carry this title of virtual digital privacy Project and program officer. You know Joshua Pesca is thanked for that word salad of it’s all nouns. It’s all it’s all one adjective. 12 nouns. Joshua, you’re, you’re out. Anyway, and then CEO doesn’t get any easier. OK. Also with us, uh we have a special guest who’s gonna give a couple of syllables. Uh, let me introduce Miles. Miles, say hello. Hi everyone, it’s Miles with Fundraise up. Thanks Tony. My pleasure. Miles is sponsoring the hub next door at Fundraise Up, so I, I thought I’d give him a little. He asked to give a shout out, so I said sure. And uh they’re giving away free socks. That’s what fundraise Up is all about socks and what else do you do at fundraise. Right, so we help nonprofits raise more money with AI and we do that by not using any identifiable information and are completely compliant across the globe. All right, that’s what a segue and not even reversal incredible. All right, you’ve overstayed your welcome. That’s enough. OK. OK. OK, thank you, Miles. No, thank you. I, he was, I, I did invite him after he pleaded. OK. So we are talking about PII. So Miles, a perfect segue, beautiful segue into personally identifiable information. Uh, Amy, we’re gonna do the overview, so I’m gonna ask Kim. Virtual digital data, virtual digital privacy project and program officer. I’m gonna ask Kim Snyder. No, I’m gonna, no, I’m hitting it hard. Uh, so for an overview, why, why do we, why do we combine these two topics? What are our issues around personally identifiable information and, uh, and artificial intelligence? Kim Snyder. So they both center on the issue of personally identifiable information. So on the one hand we’re talking about what kinds of regulations exist, how do you manage your data I’m too far away. Don’t whisper, Kim. Everybody hears you. Oh, go ahead. I’m waiting. Um, now you, you edit this, don’t count on too many edits. Oh dear, OK, alright, so, um, we’re talking about personally identifiable information which for quite a while for the last couple of NTCs have been talking about this here and. For quite a while it’s been about more about regulation this year I have to say it’s about having our data out there and vulnerability and so looking at data management and how do you start to take stock of your data so that it is less vulnerable and the person the people whose data it belongs to is also less vulnerable and the other topic which I’m here with my co-facilitator um. Uh, Shawna is with all the amens and I’m here. I’m just like I’m a man, yeah, in the, yeah, so, so talking about how that what constitutes personally identifiable information, how much that’s expanded in recent years and Shawna, what’s what’s your bright lines, how are you related to. Yeah, yeah, so Bright Lines, I founded it 4 years ago. We are a doxing prevention company for folks who don’t maybe know what doxing means. Yeah, it’s define it please. When folks will use your personal information or sensitive information, they’ll post it publicly, essentially posting your documents, that’s where doxing comes from with the intent to incite others to do you harm. So there’s like a malevolence there, right? I don’t usually consider it doxing if someone posts like. A relatively available email address from like a professional setting. I do consider it doxing when it’s your personal email address and the intent is to ask others. It could be your birthday, it could be, could be your wife’s or my man right here, yeah. the PII PII is an expanded. No, I never, no, no, actually I came out of US intelligence community. I was there as a much younger person and in a different age in the United States and in terms of our national security. It was really progressive national security person, um. The whole community, yeah, the I I’ll just say the I mean the intelligence community, yeah, yeah, I don’t usually get too granular with that um but the. Was it in the session description it would have said OK yeah we can talk about that. OK, well, I, I’m not sure I’m, I’m pretty sure, but there again it’s one thing when it’s like out on the airwaves. First is when it’s in like a session thing yeah and at at the time when I was there I was detailed out to the DEA this might have been what you read, to train them on finding their targets on the US side of the border of drug trafficking organizations so we were using these same techniques. I was training them in these like techniques to find people. We reverse engineered that now four years ago after the 2020 election when. Folks were going after Ruby Freeman and Shay Moss for just passing a piece of gum while tallying ballots in Georgia they have a penthouse in Manhattan now have the keys to that penthouses. Um, OK, interesting. So reverse engineer I see reverse engineered your, uh your prior prior work. All right. um, so referring to your session description, uh, how AI and big data are transforming privacy risks by enabling aggregation. So your concern is that the, the. Attempts at uh. Spamming people, not spamming but spoofing, fishing, they can, it can be so granular and so accurate that they, they look more and more real. This is a part of our problem, right? OK, and people and agencies, people are using artificial intelligence to gather this information and then and then put it together and collate and then threaten. So they will, so I think we could probably tag team on this. Do you wanna do the production part? So what we see is them gathering data. There’s a lot of data that’s out there about all of us, and I will. If there’s one point folks take away from me talking today in addition to my hype madness, it’s that this is not your fault. Our clients come to us and they say, oh, if I just hadn’t shared so much on public on social media publicly when I was younger and it’s like no no this had nothing to do with you. Your public records are being scraped by data brokers every day. If you own a property, if you’ve ever registered to vote someplace, if you have a driver’s license, which you have to have if you wanna get on an airplane, that data is being sold or scraped. So that’s the data that’s the source data for data brokers. So yeah, sometimes for free, for a, yep, OK, but publicly available, you don’t need to be, not an agency there’s no kind of like legal process to gather it exactly. This is why law enforcement officers, like certain law enforcement agencies now go around legal process and we’ll just buy data from data brokers. Oh, so much easier than defending a subpoena. to prove it to a judge to prove it to a judge and then if this if they move to quash the subpoena, you have to defend it. Exactly. So AI can now gather data from various sources, so it could be used to scrape these sites. It can then be used to connect data. Let me share a story. We got a phone call like a very concerned client. They had just received a phone call themselves from someone who claimed to have. Photos of theirs compromising photos from an old Snapchat account and on the call they described a photo that this that our client knew they’d taken right it was a photo of a room they were describing a room and the clients like, I remember that room. I remember that poster that they’re describing. I think I might have posted it on Instagram one point it was public, but how did they get my number? How do they know where I work and. My response was like, this is a scam. Someone scraped, someone bought a scra of LinkedIn. Maybe they connected that to your phone number. Maybe you have your phone number connected to LinkedIn because you use it from MFA for multi-factor authentication. They connected that to a handle on Instagram, probably using your face, a facial recognition. And then they just made this phone call and talked to you about your employer finding out about these photos, which was a bluff because your employer’s name is listed on your LinkedIn profile. It’s terrifying for her. And Kim has taken it a step further. So you can stitch all this together, right? and you can process all this data at speeds that never were possible before, but you can also use generative tools to create things so you can. Easily mimic a style of someone so you can also so you part of that data that you grab off of LinkedIn or social is somebody’s writing style so you can, you know, generative AI is really great tone and style and also events. So if you’re posting about events and things happening you could get. An email from your purportedly from your executive director or a colleague referencing that event and things that happened and people who were at that meeting it depends on how public the data is and then you know that can be used as a basis for a you know phishing email um that is a lot more convincing phone call yeah or a phone call this person that called our client was a human but they don’t have to be we’ve seen cases where EDs are being impersonated. And it’s video and it’s audio of them that is so convincing to the people that they’re reaching out to and this is it’s trivially easy to do right in our session in fact we had which one is the real Kim and there were two videos of me and one of them was not me um it was AI me but that cost me $29. To take that, so it’s not inaccessible. These tools used to be it used to be like really hard to do this or 25 cents and it’s like a photo in 3 seconds of audio, and they can make those videos, yeah, and you can have me say you don’t even need me saying the alphabet or or Kim’s title for Christ’s sake or half of Kim’s title. I did say you could swear. I didn’t say you could take the name of the Lord. There’s a difference. There’s a difference. There are boundaries even on nonprofit, there are boundaries. This is Chris. I’ve, uh, I’ve gotten, I’ve gotten these, uh. Dear Tony, I know I could have called you at my number or or written to you at my address accurately, uh, but I chose this method instead. So now I know they’ve got my email and my phone and my address, uh, included a picture of my home, which they probably got from Google Maps or, or right, and, uh, I, I some kind of bitcoin bitcoin scam. But how did that make you feel uh the first one I was a little like. Yeah, I was a little nervous, but, but I’ve gotten, uh, we all have gotten Bitcoin scams in the past, but this one had, like, you know, like you’re concerned that amount of information a lot of, yeah, yeah, it had the right and uh I, you know, I, I ignored it with some trepidation and then like a day or two later I got another one and you know I knew I was just kept coming. It was bullshit. Yeah, I saw one of those from one of our threat intelligence partners, someone who swims in this every day, and it terrified him and his wife. Yeah, because it’s so it’s so close to you. It’s why receiving one of those phone calls or back in the, I would say back in the day I got really energized around Gamergate started to try to support the folks who are being targeted by Gamergate. This is back in 2015, and they would describe what it was like to have like, you know, I sleep with my phone next to my bed. And or under my pillow and to have that be the stream of all of this like directed hate messages like you should kill yourself or I’m gonna do this to you or I’m going to do this to your parents or whatever the case might be. It’s so proximate that technology removes what feels like barriers between you and everyone else, and the issue with doxing so terrifying is that you don’t know who it is. It could be anybody. How do you walk down the street? How do you like sleep in your home, not terrified? You don’t know. I never thought about that. Who’s coming after you? Thank you. I never thought you bet new nightmare unlocked. Yeah, no, no, you know how, but Tony, so you get these things because you’re you’re killing me. It’s supposed to be reassuring us here on nonprofit radio. Well, you’re terrifying. We’ll get to that. We will get to that party eventually we’re we’re great parties, but, but, OK, so you’re, you know, more public person, uh, you, you know, nonprofit radio, so, so you. Get these things it’s a little unsettling and unnerving for you, right? yeah like so imagine how like a nonprofit staff person who happens to be working in an organization that may be more targeted by malicious actors, OK, so one is so your staff member starts to experience this and this may this could freak people out, right? So that’s who we’re thinking about. Um, and kind of raising the awareness, OK, yeah, I mean these are folks already dealing with some level of cortisol at a on a regular basis because of work because of their mission. I think we’ve spent enough time on motivation, and let’s let’s, uh, let’s let’s transition, uh, not subtly very abruptly to what the hell do we do? What do we do it already. Is it already too late? It’s never too late. I’m sure you’re not gonna say it’s too late. No, I wouldn’t be here. Yeah, well, I also believe it and I’ve had those moments. Listen, I live in DC and DC DC Health Link had their data leaked and taken a number of years ago and my child who had not even turned a year old had her social security number lost in that breach and I was like, oh man, she’s not a year old, you know, like how is this? This is the world we live in, right? And I turned to my partner and I was like, this is just, I don’t even know why we bother. And she’s like, you can’t, you of all people can’t have that feeling. It’s OK that you do right now, but you have to keep going. No, there are plenty of ways to ameliorate it. Yes, let’s get, let’s get into them. So what we’re with you. Why don’t we start? Go ahead and then we’ll go to Kim. Yeah, I think you can think about this so the individual as the vector to threat to the organization that can be reputational financial threats to the organization could make it hard to fundraise if you don’t support that person very well. Um, you, you would harm your reputation, say, or, um, it could make you look illegitimate to your funders, right? So if you can think about where the risks are to the organization, that’s one set of what to do, right, action items, and I might leave that with you and speak more to the personal. So when it comes to protecting yourself as an individual, there are plenty of ways that you can work to remove your data online was referring to Kim, not me. Oh yeah, no, Tony’s not gonna take that part no Kim’s got that, um, Kim. I won’t try your title um when it comes to the individual, listen, all of us have data out there again it’s not our fault we have lived a life, right? Like we’ve done things it’s, I think it’s a betrayal of trust in our own local governments that they sell this data and no one’s ever asked us for consent they’ve never informed us, etc. etc. etc. OK, so what do you do? You can sign up for one of those services that removes your data from data brokers we consider that like um. Like taking Advil, right? Like it’s like kind of taking care of some of the pain and some of the symptoms. What we also recommend is like looking back to the source data itself. So if you own a property that you live in, we always recommend that people consider moving it into a revocable trust that they don’t name for themselves. You’ve seen too many estate attorneys call it the Tony Martignetti revocable trust. Exactly exactly a different a different name to the revocable trust. That’s it. So now the ownership is obscured its data that’s already out there from prest. This is the argument that our interstate attorney always gives us and we have to educate them on this. They’ll say, oh, but it’s your name’s gonna be on the document granting it to the trust, but your name was there before on tax documents. The way data brokers work is that they’re constantly pulling this data down and renewing their data set. So when the new data comes down at this address, they want the most accurate, the most recent. they’ll overwrite it. So it may be that you lived at that address at one time but you don’t any longer and if someone’s looking for that address, it’s not your name on it. So it will get overwritten, especially over time. What we’ve seen wildly enough is that when that piece comes out, it’s like a house of cards. When you pull that property record out the rest of it tends to fall apart. We see our clients less and less on ownership is kind of a uh. a core or a hub to to other data yeah absolutely yeah I think there’s some connections happening there with like app user data that’s also on an ISP that’s connected to the house, etc. etc. is there other pieces about that location um that create profiles anything else we can do on an individual level besides the uh property ownership. Another big vector is voter data and I know that’s probably not popular in this audience because a lot of folks believe a lot in the voter file and voter data and using it and I, we often see voter data on getting used mm. Getting bought and getting scraped and so we will recommend that folks apply for programs in their states called address confidentiality programs or safe at home programs they’re always set up in with uh survivors of intimate partner violence in mind but a lot of the programs are pretty expansive, so if folks are concerned about stalking or harassment they can also apply and that then gives them a proxy address in some states like in New York across all agencies. So the DMV is now not going to sell your home address and your name. They’re going to sell your your name and your proxy address together. And and shout out the names of those programs that you would look for at your state. Address confidentiality program or safe at home. If you’re interested, the National Network to End Domestic Violence NNEDV.org has a comprehensive up to-date list of those programs. OK, awesome. Kim, uh, before we turn to Kim, uh I think you’re the perfect question perfect question answered. Person, you’re a person, you’re a person. You’re neither a question nor an answer. You’re you’re just a person with a lot of answers. Um, I read once, it’s so hard to unforget, you know, to unlearn things that, uh, the value of, of stolen data is really in the future is more financial like so that the bad actor can act without you tying it to a specific event. So my credit card, let’s say a credit card number is compromised, it’s of more value if it’s 3 years old than if it’s just a couple of weeks it was just stolen a couple weeks ago. Is that true or is that incorrect? I can see that. I can see that being true. Maybe we’ve gotten a little bit better banks and credit cards have gotten better about just reissuing new cards. Websites tend to push you to change your password when they’ve alerted you that there’s a breach, so I, I think. The private companies more so in government agencies but private companies I think have caught on to that a little bit and I think there is some truth if it’s not for financial means but really someone trying to go after you, we call that a ideologically motivated attacker. What we saw you used the word vector before I did, yeah this is my background so they um. What we found with uh a university, a client that’s a university, their students were being targeted. Some of these outside groups showed up to student houses over the summer. The students had already graduated. We’ve gotten some of their address stuff removed. The addresses weren’t available in connection to their names online any longer. So what we think happened was that those addresses that was screenshot and saved. That can happen, yeah, so it’s not a perfect fix. However, what if you have one as an intelligence officer, if you have one data point, so you have that screenshot, but then you have all these other things telling you that Shawna Dilla no longer lives at that screenshot address, you might show up there, but you’re not gonna spend a lot of time on it because you can’t verify it. You can’t confirm it with another source. Makes sense? Yes, thank you, thank you. All right, Kim, let’s turn to you on the organizational level. What, uh, what can we do, uh, there to. Protect ourselves from what’s already out there. How do we help nonprofits and small and midsize are our listeners. Alright, so for many years the the kind of mantra has been to verify, verify, verify verify. I thank you very much, that’s Kim Snyder and Shawna. No, I’m joking. She’s like I’m we’re out of time. No, we’re out of time. Are we out of time? No, I’m only child I fall for jokes very easily. I wish I had known. I wish I had so many. I had so many more. I had so many more in mind for you specifically talking about a targeted attack. Oh my, talk about a vector vector I was coming right at you. I could have written that you’re you’re putting this on the airwaves. You know how vulnerable you are. Oh man, I got all kinds of advantages. All right, I’m sorry, I interrupted you. What was I talking about dying. Go ahead. OK I’m sorry. OK, so we used to talk in cybersecurity world about, you know, verification verify, verify, verify that was the mantra, right? So now we kind of reshape that so that it’s vet and verify so have kind of multiple ways of verifying especially incoming requests. Anything kind of trust your spider sense is what I’d say if something seems a little bit off like what what are we talking about? So if you receive an email, if an email comes and it, you know, it comes from your development director who’s saying who’s referencing something that you just went to the panel or if it comes from accounting, write a check if any money is involved. And it wasn’t like completely expected even if it was a little expected actually I’ve seen I’ve seen this happen where people got into um nonprofit systems and using AI can scan what’s going on very quickly. And then target things that are about to happen from kind of things that are OK, so, so I would, so the instinct instinct, OK, use your, use your instinct but also make it a policy, make it a process that you just follow uncomplicated process for verifying like any financial transaction needs to be verified even if it’s expected, yeah, so yeah, so you wanna walk through that. You just get much, much more deliberate. About verification and and who is it coming from and you don’t want to. Confirming, did you send this email or not replying to the email, but my phone yeah exactly yeah you you send this email about this rush transaction or or routine transaction. Do it in a different format right different channel, yeah, so you know, and even though the instinct may be email back quickly but no right um but then what you do also is create a culture in your organization where that’s OK to do where it’s OK to take that extra 30 seconds minute to you know verify to ask someone for their time to say I just wanna check, did you send this to me? Um, and in that way it’s OK even if it’s because he’s actually director you can say, did you send this to me? I just wanna make sure and so that that’s an OK thing to do. In fact, that’s a good thing to do. Now we can’t they have to be boundaries around this because we can’t do it for every, every message we get so you mentioned. financial financial transactions and no no no not nervous at all financial no no no financial transactions, any kind of initiated correspondence where they’re asking you for something or for some information. I saw a scam recently where the uh an an old employee was trying to be reinstated and wanted to go around HR to IT to get their accounts reset up like I’m I’m coming back and it was like using the person’s middle name so it’s already a little bit fishy but. They went all the way up to the CTO of the of the company and said hey so and so and these people were friends on LinkedIn and like had shared messages back and forth so the attacker knew this was a personal relationship. hey so and so I’m trying to get reinstated. They’re telling me you need to go to HR, but like I but I can do this. I just need to get my account access back up and online and the CTO is like no. Oh bro, you gotta go through HR. I can’t do anything because they had those controls in place, but small and let’s be fair, small and medium sized organizations don’t, so I’ll just take care of it now or we don’t have a, we don’t have a we don’t have any clear guidelines that we give to people for all requests we need to go to HR. I thought of another. Potentially nefarious request you send your logo. Could you, could you, I need a I need a high def for the logo, you know, the, the, the, the JPEG I have is, is not good. I need a high definition logo that could be that could be to produce a check that could be to make a spoof a spare a spoof website, um, OK, I mean, but it seems innocuous send a logo, yeah, it’s very easy to spoof a website, right? So you know, you know, check. Also check where it’s coming from, right? So you know I’ve had an organization where there were two spoofed, um, there’s spoofs on both ends a spoof of the funder, a spoof of the the grantee. Can you tell us more about that story? It’s a really good one. So yeah, so they, they got into an organization’s, um, you know, Microsoft environment. I asked the questions here whoops. Go ahead. Uh oh, off the mic. 3 like 30, go ahead. So, um, Anyway, that’s late in the day. And I’m thirsty. Yeah, late in the day it’s not it’s, it’s well it’s almost 3 o’clock. You’ve been going since then nonstop. Um, anyway, all right. So the organization had someone get into their systems for a very short time, but in that short time they were able to tease out some information again this is AI can help with this kind of analysis short you know canal is a lot of data that it can grab very quickly and um identified some upcoming financial transactions which were rather large and so um in order to kind of trick. The person to sending to the wrong place, they set up fake websites, fake websites for the foundation, fake websites for the grantee, and domains not websites domains, and so then they had emails coming back and forth you could hardly see the difference and so the, the, the real people, the real people were communicating with the bad actor on both sides and the money. And he got sent to the wrong place, OK. Yeah, that was, that was actually no they did great, but, but it was that was a happy ending, but not necessarily. We started with Shawna, so we’re gonna end with Kim. give us oh no we did OK well I’m not Shawna, your mic is down but that she still gets through. She talks and laughs so loud you hear her over Kim’s mic. No, I didn’t, I did not but one more thing before, before we unless we’re totally out of time, um, don’t shoot the messenger. So create a culture. This is another thing that’s any size nonprofit can do where if something happens, if you click on that thing, if you did that thing that you feel like uh. That was really dumb, right? Make it OK to report that and you don’t get in trouble and there’s no shame and blame because it happens so but yeah the the no blame kind of we encourage you to. You know, say it, yeah, call yourself out, yeah, and there’s no punishment, you know, some organizations like they don’t want bad news at the top, so. All right, we’re gonna leave it there, OK? All right. That’s Kim Snyder. Virtual digital privacy project and program officer Roundtable Technology and Shana Dela Vu, CEO CEO Bright lines. Thank you, Kim. Thank you, Shawna. It’s a pleasure. Shawna laughed her ass off. I’m a good sense of humor. All right, I love it. Uh, and thank you for being with a, uh, well, whimsical, I’m not sure it covers it. Raucous maybe, uh, at one point, uh, uh, uh, anarchical because, uh, there was a question that I did not answer. Uh, session. Uh, thank you for being with us at uh 25 NTC for this episode sponsored by Heller Consulting. Technology services for nonprofits, virtual digital privacy project and program officers. It’s time for Tony’s Take-2. Thank you, Kate. A new tales from the gym episode just happened this morning, this very morning. I was minding my own business as I do on the elliptical. And overheard two women talking. One lives here permanently, and the other one who said her name. Sandra Lynn, uh, she lives in North Carolina, but not here in Emerald Isle. She lives, uh. In the Raleigh area, like that’s about 3.5 hours, 4 hours away, roughly. And she was lamenting, Sandra Lan was that uh that she can’t live here full time, house prices are high. And she also still has, uh, her mother and her father-in-law, so her husband’s father are still both alive, and so she needs to stay in that area, but she was, you know, looking forward to retiring here sometime but lamenting that she couldn’t live here now. And that got me thinking as I was on my. 6th or 7th uh interval on the elliptical. I do 88 episode 8, Not episodes. What did I just say? 8 intervals. I do 7 intervals of a minute, take a minute in between, and then the last interval is 2.5 minutes. I was toward the end and it got me thinking, listening to Sandra Lynn. That, uh, I’m grateful that I do live here full time, permanent. This is my home. And that, you know, it’s that there are other people who don’t live here who wish they could, you know, so, uh, you know, I, I add, I have, I have a long list of gratitudes, but I don’t specifically say grateful that I live here in Emerald Isle full time. So I’m gonna add that to my gratitudes that I do every, I guess I’ve told you every 2-3 times a week. I’m adding. Gratitude that I live here in Emerald Isle full time in this beautiful place and I have the ocean across the street. Uh, your own gratitudes. I hope you’re, I hope you’re doing your gratitudes out loud, at least a couple of times a week. That is Tony’s take too. Kate. You do sets. Uh, well, sets are for, yeah, no, that’s different intervals. Intervals on an elliptical, you do a minute hard and then a minute resting. And then a minute hard and a minute resting, it’s called high intensity interval training, HIIT high intensity. It just means you do intervals of things like you sprint, yeah, I don’t run, I’m on elliptical, but you might sprint and then walk, and then sprint and then walk and sprint and walk. Those are called intervals. Sets are like you do 3 sets of 10 if you’re, if you’re on a weight machine or something like that, or maybe pushups, might be 3 sets of 10 or something like that. I don’t know, they seem, there seems to be a different, well, I think the interval is because you’re still active, you’re just resting in between the high intensity intervals. Gotcha. That makes sense? Yes, and I am grateful that you have a beach house. Yeah, because you get to, yeah, you get to visit and uh laze around and uh. What is the word I’m looking for, uh, not schmooze, but, uh, you get to, uh, I don’t know. I can pretend that it’s my beach house. Yeah. You can for a week, yes, but then, then I’m very happy to say goodbye. After a week. Love you too. We’ve got bou but loads more time. Here is balance AI ethics and innovation. Hello and welcome to Tony Martignetti nonprofit Radio coverage of 25 NTC, the 2025 nonprofit Technology Conference, where our coverage is sponsored by Heller Consulting technology services for nonprofits. With me now is Gozi Egwanu. Gozi is director of programs at the Technology Association of Grant Makers. Gozi, welcome to nonprofit Radio. Awesome. Thank you for having me, Tony. Pleasure. You’re welcome. Your session is AI strategy for nonprofits, navigate ethics and innovation. We have plenty of time together, but can you give me a high level view of the the topic and the session that you did? Sure. So the session was really, um, and was really spearheaded by Beth Cantor, uh, and it basically provides uh a balcony view of where we are in the sector in terms of AI adoption, ethical responsible AI adoption, the nonprofit and philanthropy sector. And so, uh, we really start with what we found in the Technology Association of Grantmakers state of Philanthropy tech survey that we did in 2024. In that survey we found what many grant makers are currently doing with AI as far as you know are they testing are they experimenting? Has anyone rolled it out enterprise level, which is, you know, at the organization wide level and what we found is that. And which mirrors quite what we’re seeing in the nonprofit world is that most folks are not using AI in terms of, you know, anything that’s crazy, you know, innovative at this moment it’s really just kind of, you know, meeting summaries, you know, taking notes, that sort of thing, um, and so and but in addition to that we found that while 81% of folks are using AI, uh, sorry, while, uh, oh sorry, 81% are using AI but only 30% have AI use policies, so. You’re using it but you don’t have any guard rails you have no way to tell your teams or your staff, hey, this is what we don’t put into the AI this is what we do put in so you’re really running the risk of having your information potentially used in a way or trained uh an AI model that, um, you know, could potentially put your members at risk, your grantees at risk, whatever the case is for your organization and so. With that little bit of an overview it basically came down to the importance of AI experimentation and really do starting slow starting at the very base level working with your teams to kind of talk through should we use AI if we did use AI what would that be for? So thinking about the use cases, the business, um, the business use like what what would be the business case for it and then you know assembling a nice team of folks, you know, as advisers or experimenters and champions at your organization. Uh, to really kind of help you all start doing that experimentation in a safe and low kind of like low risk way, um, and then from there really defining whether or not AI is your, your next move and then once you do have decide that AI is the next move you wanna move into that next level of the AI maturity which Beth, you know, covers really um really well uh you know you go from that exploration to discovery and then you move into experimentation and ultimately enterprise eventually. Um, but what we’re finding is that most folks are not there yet. They’re still very much experimentation early stage, very early stage, um, and, uh, you get to kind of get to see a case study of it through the work that Lawan did at her organization United Way Worldwide. OK, well, we don’t have with us, but you can provide a lot of context, lot of, lot of detail, I just said you could talk. All right, um, are, are we, do you know the you might not be part of what you surveyed, but was there even intentionality around should we, should the should we use question or did it just kinda happen because people started, people started hearing about it using chat GPT. Well, you know, with one of the questions that we did on the survey, we found that like there’s quite a few folks that are using it in what we call shadow use or shadow AI, which is basically you’re using AI but your organization doesn’t know what you’re using. I see. Alright, so that’s not intentionality at the organization level. No, no, no, I would say not, not. Uh yeah, so we really want to encourage the intentionality which is don’t start using the AI unless you all have that collective organizational conversation of is this something that we should be doing? Is it useful? Is there a business case to go with it? Is it relevant? Does it make sense? Is it safe for our organization? does it align with our ethics? And then consider going into experiments. OK, let’s explore that question a little bit uh now in 2025 because I, I suspect at 26 NTC we won’t be asking the threshold question, should we, should we use? So what, what, what belongs in the conversation if we’re, if, uh if we’re at the stage where Well, uh, individuals may be using it, but we don’t know. Or if nobody’s using it and we’re trying to decide enterprise wide, you know, is there not, we’re not even at the is there a use case like but should we, should we explore it? What goes into that conversation? Sure, um. Again that you know, really thinking about the business case. So when you’re having that conversation about should we use AI, then you have to think about what would be the specific usage of it, right? So say you’re the finance team and you’re considering using AI, what would be the benefit of using AI versus doing the doing the the work flow or process that you currently have and you’re thinking of having AI do? so you really. Kind of have to have that conversation like an in-depth conversation about the process that you’re doing right now. Is there anything wrong with it? Are we losing anything? Could we gain, uh, productivity, time in our days and our schedules if we were to move to using AI to do this one process or this one, this one work flow? Then at that point you think about, OK, maybe we do get a benefit out of it now that we get a benefit out of it. What are some of the things that we have to be concerned about now that we have a benefit is it that now we don’t wanna make sure we wanna make sure that any financial information that could be sensitive to any of our donors or their their personal information, do we not want to have that being able to be, you know, used in the AI model or whatever system that we’re using so you know, you, you start with here’s how we do. Things here’s how AI could potentially benefit and then you move into that conversation. OK, if we did, what are some of the risks and concerns really thinking through all of them as much as you can, we know that you can’t think for every single possibility, but as much as you can kind of write it out and map it out as a group with several folks in the room, the better that you are at being able to say yes or no on moving on with AI as that. Potential new solution. OK, and a part of what goes into this intentionality is a usage, a use policy, your, your, you know, you want us to be thinking about ethical uses. OK, uh, what, what are the, what are, what are the ethical concerns? How can you, how can we talk through those? Well, you know, one of the key ethical concerns is that we know that most AI models that exist now, including open AI, were trained on the internet, and we know the internet can be, uh, wildly biased, wildly biased, filled with lots of terrible things. Not only biased but misinformed, misinformed wrong yeah complete nonsense in a lot of cases, um, and so if you’re using these open AI sources that have been trained on the internet, then you have to be really careful about deciding to use it against, say your theory of change. So if you’re an organization that is er. Be uh vulnerable populations groups that are already kind of under attack, whatever the case is, do you want to have AI making or informing your decisions related to work that you’re doing with these vulnerable groups? More than likely no because the AI may choose to do things that are more in line with the group that is. Biased that may have you know may be unethical and so you want to make sure that whatever you’re using the AI to do that it isn’t putting the organizations and the people that you support and serve in harm’s way so really thinking through, hey, if we’re gonna use it in this way, maybe we need to use it in a way that does not put these groups in harm. Maybe we just focus on using it internally like folks do for the meeting. Notes because that’s a very low risk thing whereas if you’re you know input you know uh decisions about whether or not to continue funding an organization or trying to measure or not whether or not their impact is aligning with your organization’s missions and values some of those those questions are not as clear cut as yes or no, whereas an AI that is trained on purely just wanting to see impact, purely wanting to see a return on investment, which is not always the case of what happens in philanthropy. Then you really have to take, take a step back and say is this the most ethical decision to go forward? Could we be putting organizations in harm? Now you can control what a model is trained on, yes, but that requires something proprietary, right? You have, you have to pay a developer to, uh, to create that. I get I don’t know it’s called a small language model. I don’t know what it’s called, but something that’s trained only on your own data, but your own website, maybe your own documents that you that you provided, but that, that requires a fee and a and a developer. Exactly, it it can it can cost, it can be expensive. The other option is if you don’t want to go the route of creating your own AI you do a paid version because we know the free versions of AI specifically I’ll talk about open AI there’s not a whole lot of freedom or flexibility in turning off the settings to prevent it from training the model on the data that you input. And so in that case you definitely need a use policy because some folks would probably just be like I really need to you know analyze all of this data on all of the groups that we served in this, you know, community that is already really, you know, under attack or potentially in in harm’s way and then now you’re putting that information into the AI to have it, you know, into the free AI to start doing it’s now. and now the AI has all of these people’s information and can now use it to provide it to other people who may look them up or want to find data on. That’s you’ve you’ve shared data that it’s gone. I mean it’s yeah yeah yeah there’s no control. So yes, enormous intentionality, care, um. And what if we don’t have a, you know, we don’t have a, a chief technology officer, chief information officer, you know, it’s an executive director, CEO, and, and maybe decent sized staff. I don’t know, 35, 40 people, but they still don’t have a chief technology officer. How do we, how do we uh ensure the intentionality and care that you’re, that you want us to? Yes, um, there’s a couple of ways, and I think oh good, I think at the core of it you don’t have to have a CTO and even yourself you don’t have to be a technologist. I would never classify myself as a technologist, but we can, there’s ways to find training. There’s plenty of training and 10 it has fantastic training for AI certifications for professionals in in the nonprofit sector, um, and I’d love to share that and 10 and tag are teaming up and we will be offering one for philanthropy professionals very soon. And so these are opportunities, a very, you know, relatively easy ways for people who don’t have that technical background to learn about the AI itself, get themselves familiar familiarized with, you know, what they need to be doing to protect themselves and their staff, ways that they can start to experiment in a safe, you know, safe space, um, so and there’s plenty of also free tools, free education. I will, you know, even I, even though I’ve talked. About OpenAI a lot. OpenAI just announced their OpenAI Academy which has all free resources and tools for learning how to utilize AI for anyone and so there are plenty of free resources out there and people online, you know, uh, there’s plenty of folks on LinkedIn that I see on a regular basis that are sharing information and providing some guidance for nonprofit leaders as well as, uh, folks. That are just not technically inclined so there’s ways that you can kind of upskill and train yourself to understand how to use AI even if you don’t have that technical experience in house. Say a little more about this partnership, can you uh and it’s technical association of grant pardon mechology Association of grants thank you um. Yeah, so I don’t have a whole lot of details to share, but essentially if you’ve, if you’ve used any of the great training and certification resources on the N10 website, we are essentially trying to make a parallel version of that same professional certification for nonprofit leaders using AI for. Our foundation leaders and so uh you can expect really a kind of a similar learning process but however it’ll be tailored to some of the different functions and needs that we find at the philanthropy you know at foundations versus what you would see at a traditional nonprofit. OK, so I’m sorry, it’s intended for professionals I should say. Um, Alright, what, so thank you. You know, that’s important ethical considerations, um, anything more on ethics because, uh, then I I want to talk about the policy, what belongs in your use policy, but is there more about ethical concerns? OK, OK, OK, enormous. I mean if you, if, if you’re exposing your data. And, and it’s gone. It’s, it’s out there like you said, right, um, our use policy that, uh, only 13, 30% have, although 80% are using AI. What goes into this use policy? The use policy essentially just outlines what you and your team should be thinking about before you ever use any AI, so. It’s kind of that no go or go kind of conversation so if it’s sensitive data, if it’s information related to any of your members that you just wouldn’t want anyone to have outside of your organizational members probably wouldn’t want to put it into an AI system so it just kind of outlines, you know, essentially guardrails for for teams and and staff to understand how to best utilize it. And I think some folks are also, you know, thinking about the environmental impacts of using AI are really now making sure that their data use policy or the AI policies are also, you know, having folks be ethical about how they’re using when they’re using AI right? so you know if it’s to do something that could take you probably about the same time that the AI does, don’t use the AI um if you’re just, you know, just tossing anything, any old thing and they’re asking questions all day probably also not a very useful. Use good use of AI you really wanna think about AI very strategically and intentionally, right? You wanna make sure that if you’re going to the AI, it’s for something that you know it’s gonna save you significant amounts of time. One of the things that I often will use AI for is drafting, you know, large descriptions for events. That takes me sometimes hours if I give it to AI, I can do it for me in seconds and the key to descriptions of events, yes, like, so we have webinars events that we have on our website, yeah, so you know I, I, I, I don’t wanna sit there talking about all the learning that you’re gonna get out of it and the objectives and this and that and so AI, I’ve trained, I have like a GPT that is based on kind of like my voice that I provide it like hey here’s the prompt, here’s what I’m kind of looking for. It provides me a draft and then I use that draft and I manipulate it how I want. Um, and so you really wanna make sure that you know when you’re prompting the AI or you’re using the AI, it’s they’ve measured it. I think one prompt uses as much energy. I think it’s like an entire city like it’s crazy. It’s like like it, I, I don’t use my quote me on that, but it’s enormous. There’s quite a bit of energy, and I can actually actually share a link to um one of the stats that came out about it. There’s a researcher that’s been sharing a lot about it, um, and she was just interviewed by, uh, I believe it was Doctor Joy Bullumwini on, uh, by the, um, the. AI justice uh group that she she leads, um, and so there’s a lot of it there’s a lot of energy being used so if you’re gonna use it, you wanna make sure that it’s for something that you don’t need to, you wanna learn prompting good prompting, so you can get what you need out of it and then you can make, you can, you know, refine it and make it better. Sometimes you may have to go back in and ask the AI to refine, you know, what it did, but you really do wanna keep it to a minimum. You don’t wanna be using AI. Constantly because the energy use and the impact on the environment is extreme extreme that gets over to the ethical concerns as well exactly because it’s yeah so yeah you’re you’re just really um basically telling your teams here’s the here’s what we expect out of you when you’re using AI and these are the things that could result in consequences if you don’t follow this policy OK um. What else, anything more about the policy, what, what, what belongs in there? Um, You know, I think the the key things is like you know making your team’s aware of the types of AI that are provisioned because that’s another thing some organizations have taken the decision to block certain AIs that they don’t want you using or even turning off certain AI functions in their uh current tech stack. So, uh, you wanna make sure that it’s really outlined very clearly the types of AI that are in use and also it may, you may wanna include something in there about how you, uh, communicate your use of AI to your teams or other people outside of your organization so. Kind of a, a nice, nice little bucket of what’s internal external, and then also where can you go if AI and where should you not go disclosures to the public um why would there be some uh some platforms or that are that are ruled out? Well, because You know, one of the things that I’ve seen some members talking about within, you know, the tag space is that there are some AI that do not allow you or some systems that do not allow you to turn off the AI function meaning that you don’t have any control of how that AI is taking your data that you have in that tech stack or that tech tool. Oh, you don’t have control no yeah and in in fact there was actually a conversation about a specifically a DAF uh platform that actually. Made this clear to many many many of our members who use it and so that is something that you really have to be concerned about is do you have any level of control if you don’t have any level of control and how the AI is using your data in that system there are organizations that would likely say this is a this is not a system that we would allow you to use. OK, it’s a good example. Um what else uh came out of the session? We still have a couple more minutes together. What else did you talk about in the session that uh that you can share with us? You know, one of the great things that we did was we did these scenarios, um, that Beth Beth put together about, you know, what are some of the things that you would say if you’re in a situation when where, you know, say for instance, uh, your organization is really excited about using AI they wanna jump head first and they just wanna start using AI without, you know, and and they they basically just want you to start rolling it out and get your teams on board. Um, and so in that scenario we really talked through all of the processes, you know, first of all, that first conversation that we talked about, like, should we even use AI that didn’t happen, so that needed to happen. The other part is also, you know, how fast do we wanna roll things out? What are some of the different change management principles that we should be thinking about as a team that could make AI adoption more beneficial and successful so really, you know, starting slow but really starting at the very beginning of like should we or should we not like that should be your because truthfully many organizations do not need AI. It’s true. I mean, it’s just the reality. Some organizations will never probably need to use AI, and then there’s a whole lot of them that probably will. So that question of like, should we do it has to happen first, um, and I think if you’re doing it on your own as a rogue, stop, do it on your own time. You want to practice on it, do it after after hours on a weekend. Exactly, exactly, not on our computers, not on our sisters. Yeah, yeah, if you, and that’s actually one of the things that, um, you know, we’ve seen a lot of our members and foundations, and I think Beth has also seen with, you know, some of the work she’s done in the in the sector is that a lot of foundations are now trying to just get to the staff and say, hey, look, we know that you’re using, can you just tell us and try to make that trust, build that trust with each other and I think that’s gonna be really a good way to help prevent a lot of the issues. Alright, let us know, but then stop. No, there’s no repercussion for reporting yourself, but only, well, only after what you report after the report date, you’re liable. All right, stop it. Exactly. OK, going rogue. All right, um, anything else? Uh oh, questions, any, uh, provocative or memorable questions that came. From the audience I’m trying to think. Um, No, well, you know, the one that had come up was just, uh, you know, there was a, there was someone at the front that had asked about, you know, AI hallucinates, and I was, and, you know, should you hallucinates, yeah, and she and the, the person was basically saying, you know, be careful using it as an organization because it could give you answers that are just factually wrong and so you know our response was like yeah you’re right AI does hallucinate but that’s why it’s incredibly important and I. And I didn’t even say this myself, but at the beginning, which is if you use AI, you always wanna make sure that it’s for something that you have a certain level or high level of expertise or knowledge about. So you know if I’m asking you to write descriptions for me, I know about the event details so that I’m not just gonna let the AI write a description and let it go and put it on the website. Yeah, that sounds good. I’m gonna put it no you review it, you make sure that. The details it’s including are correct. If there’s any statistics or numbers that are being used, you can go and verify those data. So if you’re ever using AI for anything, you should always have a human in the loop. There should be someone that’s able to verify the information, especially if you’re someone that’s not knowledgeable in that specific thing that you ask AI to do. You need someone who is either that or it’s gonna be written at such a high level that it’s maybe that has no value. Exactly, exactly. All right, how about we leave, are you OK leaving it there? Oh, you feel like we covered this? I think we did. OK. All right. All right. Go the Abuno. Euanu Gozi Ebo. Director of programs at Technology Association of Grant Makers. Gozi, thank you very much for sharing all that. Thank you for having me, Tony. My pleasure and thank you for being with Tony Martignetti nonprofit radio coverage of 25 NTC where we are sponsored by Heller Consulting technology services for nonprofits. Next week, 225 NTC conversations to help your fundraising events. If you missed any part of this week’s show, I beseech you. Find it at Tony Martignetti.com. And now the donor box is gone, I miss our alliteration fast, flexible, friendly fundraising forms. Uh, I miss that. All right, well, I am grateful to Donor Box though for 2 years of sponsorship, very grateful, grateful. There’s another gratitude. I’m grateful to Donor Box. Now that they’re not a sponsor anymore, I’m grateful to them. No, I, I’ve been grateful. I just haven’t said it. OK. Our creative producer is Claire Meyerhoff. I’m your associate producer Kate Martignetti. The show social media is by Susan Chavez. Mark Silverman is our web guy, and this music is by Scott Stein. Thank you for that affirmation, Scotty. Be with us next week for nonprofit radio, big nonprofit ideas for the other 95%. Go out and be great.

Nonprofit Radio for April 28, 2025: #25NTC & The Human Factors Driving Your CRM Success

Amy Sample Ward#25NTC

We launch our 25NTC coverage with the CEO of NTEN, which hosts the Nonprofit Technology Conference, sharing the numbers and the experience of this year’s Conference, earlier this month in Baltimore, Maryland. They’re Amy Sample Ward, and they’re also Nonprofit Radio’s technology contributor.

 

Rubin Singh: The Human Factors Driving Your CRM Success

Don’t blame your tech first when it feels like your CRM database is letting you down. Human beings, the tech users, have responsibilities that precede, and must align with, your technology. Rubin Singh returns to enlighten us about business processes; inclusive design; personal and professional growth; and more human factors that impact the success of your CRM database. He’s founder and CEO of OneTenth Consulting.

 

Listen to the podcast

Get Nonprofit Radio insider alerts

 

I love our sponsor!

Donorbox: Powerful fundraising features made refreshingly easy.

Apple Podcast button

 

 

 

We’re the #1 Podcast for Nonprofits, With 13,000+ Weekly Listeners

Board relations. Fundraising. Volunteer management. Prospect research. Legal compliance. Accounting. Finance. Investments. Donor relations. Public relations. Marketing. Technology. Social media.

Every nonprofit struggles with these issues. Big nonprofits hire experts. The other 95% listen to Tony Martignetti Nonprofit Radio. Trusted experts and leading thinkers join me each week to tackle the tough issues. If you have big dreams but a small budget, you have a home at Tony Martignetti Nonprofit Radio.
View Full Transcript

And welcome to Tony Martignetti Nonprofit Radio, big nonprofit ideas for the other 95%. I’m your aptly named host and the podfather of your favorite hebdominal podcast. We’re kicking off our 25 NTC coverage this week. These two segments are both from 25 NTC. It was a wonderful conference. I think the best. Uh, this was, I believe, the 11th year that I’ve hosted nonprofit radio, uh, in a studio at the nonprofit technology conference, and I think this was the best one. You’ll hear Amy and I talk about that. So excited, legitimately, you know, some people say, uh, I’m excited. No, I’m excited that we are launching, inaugurating, kicking off our 25 NTC coverage this week. Oh, I’m glad you’re with us. I’d suffer the effects of ramidenia if you pained me with the idea that you missed this week’s show. Here’s our associate producer Kate with what’s coming. Hey Tony, this week it’s 25 NTC. The CEO of N10, which hosts the nonprofit technology conference, shares the numbers and the experience of this year’s nonprofit technology conference last week in Baltimore, Maryland. They are Amy Sample Ward, and they are also nonprofit Radio’s technology contributor. Then the human factors driving your CRM success. Don’t blame your tech first when it feels like your CRM database is letting you down. Human beings, the tech users, have responsibilities that proceed and must align with your technology. Rubin Singh returns to enlighten us about business processes, inclusive design, personal and professional growth, and more human factors that impact the success of your CRM database. He’s founder and CEO of 10th Consulting. On Tony’s take 2. Tales from the gym. Meet Roy. We’re sponsored by DonorBox. Outdated donation forms blocking your supporters’ generosity. DonorBox, fast, flexible, and friendly fundraising forms for your nonprofit, DonorBox.org. Here is 25 NTC. Hello and welcome to Tony Martignetti nonprofit Radio coverage. Oh wait, I should do what Amy loves. Hello and welcome to Tony Martignetti Nonprofit Radio, big nonprofit ideas for the other 95%. I’m your aptly named host and the podfather of your favorite hebdominal podcast. What a genuine pleasure to welcome. The CEO of N10, the host of the 2025 nonprofit technology conference, Amy Sample Ward, welcome. Thank you. I don’t know that I’ve ever gotten to see in person witness live the Podfather intro. So yeah, exactly, it doesn’t have the same power, you know. Uh, so, uh, we’re here at 25 NTC. We’re at the uh Baltimore Convention Center. Oh, I, I should have said that our coverage here is sponsored by Heller Consulting technology services for nonprofits. Very grateful to Heller Consulting. How’s the conference going? CEO. It’s going great from community perspective, you know, I think we didn’t really know what to expect in this moment right from the community perspective, yes, I mean you know there’s always special things um there’s always opportunities to continue refining. Uh, but you know, from the, from the community side and kind of what we spend so much time focusing on, you know, the community experience, we just, you know, would anyone actually. Feel like coming when they woke up Wednesday morning would be, you know, like are is anyone even gonna be in a space to have conversations or wanna go to a session, you know, and we’re not trying to pretend that everything is fine and normal or that those things exist, you know. I like we’re certainly embracing that and yet we’re trying to embrace it within a reality of well we’ve all planned for this conference to be here so so it is still a a structured thing and and how porous can we make it in real time together to also meet whatever needs are emerging by maybe going to a session that was already planned and you know the speaker had prepared but from that conversation came something. Oh my gosh, we, we need a space, we need to keep talking about this right now and what do we do, right? So I’ve heard a couple of stories like that anecdotally that I’ve got I’ve learned so much I need to learn more now and you realized what I didn’t know and I I need to connect with that person or, you know, um, I heard it from an audience, uh, like talking to somebody who was sitting next to the person who was talking to me and then also to this for. Speaker, I, you know, I need to connect with, I forget whether it’s him or her or whatever. It doesn’t matter. Um, I need to learn more from this person. Alright, let’s let’s, uh, you know, I always like to ask you the numbers you know how many folks are with us here in real life in Baltimore, Maryland. I don’t know that I actually have the accurate numbers. We, we’re. Uh, 1800 registrants overall, but given how many shifts are happening, I don’t know exactly, and, well, I’ll finish my first sentence and then I’ll add a second so I don’t know how many folks are necessarily in person because we don’t. Require if you’re in person you have access to the virtual and so folks that couldn’t come for whatever reason didn’t necessarily have to tell us that so we don’t totally know or they could have even come one day and then gone virtual another day that’s right so I don’t totally know how many folks are in the room. I know that I think catering told us that over 1600. Silverware were used at lunch today. So just for lunch today I guess that yeah I know I only used one and I used one upstairs even so it’s a proxy for right so that’s how many people ate lunch I guess. OK, um, talking about the conference experience, uh, we’re gonna bring in someone who was a previous guest Aia Aria Ma, come on, come on, Arya. you can share my mic, just share my mic. Yeah, let’s make it easier. Um, because this is her first NTC, and I, I, she was saying, she was saying things that I think you would want to hear as CEO. So I said, if you want to come back, well, it’s not gonna be quite that long, but, uh, yeah, it’s not gonna be quite that long, but thank you. Uh, if you wanna talk to the CEO, let Amy know that, that, uh, what, how you feel about your first NTC. So this is Ama. Um, her company, uh, are, are you just, you say your company, so I don’t have to look back two pages. OK, I’m, I’m the principal consultant founder of Lunara, so I do consulting with environmental conservation nonprofits. Um, first of all, thank you. This has been amazing. I’m actually from Baltimore, so just jot down from Boston, do a quick family trip, and the people here, I feel. Like NDC really has curated an amazing group of people where it’s not really about networking but really connecting and knowledge sharing the accessibility here it really feels like the staff are here looking out for the for the participants and it’s just been an amazing time connecting with people the bird like the feathers of a bird table conversations, the comfy chairs like this is I’ve been to a lot of conferences. And it’s definitely one of a kind. So thank you so much for curating an amazing team, curating amazing people who come here. It’s been a really great time for me. All right, thank you so much for being thank you for being one of those amazing people that is here, right? Thank you. I so appreciate that. Of course, of course, lots more NTCs in your future. OK, good. Thank you. Thanks for coming back. Glad you did. I, yeah, a little treat for you. I knew I knew you would want to hear something. Thank you. You’re welcome. Um, So, oh, I turned up for that was turned it up for our, yeah, OK, um, alright, so we have about 1800 people. Well let’s call it 1800 between friends, just as a round number. That’s good. OK yeah, we, I mean, I think there were like 12 people who just showed up and registered on site yesterday, so. Yeah, the number is a moving. When we talked a couple of weeks ago, you said people just show up. I was amazed at that. OK, it happens. OK. There’s always room at the NTC, you know. Yeah, well, we’ll add another chair. What did Ari just say? There’s a staff. Thank you for the staff, looking, looking out for all of us. Yes, of course. We’re accommodating. We’re, we, I’m a member, so I’m not staff, but I’m, I’m part of the N10 the N10 community. I don’t want to call it the the NTC people call you NTC. Yeah, I know Amy. She runs NTC in a way they, they, yeah, um. The commons experiment, yeah, I mean, I hope like we could talk about it for a minute. I’m just gonna preempt your intro and say for people who run conferences as part of the larger work of your organization, right, not that not that you’re an event planner, but people who are listening to the show and have organizations that have conferences as a part of your programming. Would love to learn what you are testing because we, as I said the other day you know there’s that like analogy or whatever like oh I pulled the band-aid off and just you know tried something we we found every band-aid and we pulled them all off at once and tested everything is is different this year. There is not an exhibit hall with pipe and drape and 10 by 10 squares that you have to walk through, uh, hallway hallways of corridors of, uh, it’s an open space, open plan. The studio here is set up right behind all the chairs that are facing the stage where the main stage where all the keynotes are and the awards were given. Um, we’re by the food station. Well, the food is here. It’s just the the hall is open. There’s not anything dividing us. Yeah, we sold no exhibit booth packages. Uh, and I admire it, I admire the attempt at change. Look, even if you, I’m not, you’re gonna decide as a team what you’re gonna do, but even if you went back to 10 by 10 cubes, uh, pipe and drape next year, I would still admire the 2025 experiment because you are trying something that radically different. You’re not bound by what every other conference does and what you. Uh, what N10 has done year after year after year for 24 years, this is our 25th, your 25th NTC. So you’re not bound by, by your own history even. I just, I admire the outward look, the fresh look, even if you go back to the way it was last year, I still will never stop admiring what you did this year. Thank you for saying that. I mean, I think. It in some ways was was and is a huge risk to say we don’t we’re not even selling packages that would equate to hundreds of thousands of dollars of revenue um so if anyone’s listening and would like to write a check, please let me know um. But it just didn’t feel like that hard of a choice for us because it feels like every year all we want is to get closer and closer to an experience for as many people as possible to be in community and we, you know, getting rid of the booths was just one thing there’s also, you know, sessions are working in different ways and there’s. More furniture in this room that is not from the convention center, you know, it’s not rounds with chairs. We actually rented every piece of furniture from a local furniture company, um, just to, I can say the name that’s my show, not the Freeman Company. We’re not we’re not using that. We actually have put on an entire, you know, 1800 person conference in a convention center without a decorator, um, so we. Did everything ourselves so that we could control and make it just how we wanted it to be. You rented how many chairs you found a source for all these, all the everything, yes, every, every vendor is local, every maker is local. Yeah, you have a little shop. There’s there’s a market, local, local vendors that yesterday I saw cutting boards, uh, I see art, uh, that’s all I saw, but there’s a market over there, local, local vendors love it. You, you also do something smart that I learned, uh, one of your team members told me. That uh you you have your staff retreat in the city where or one of your annual retreats in the city where next year’s NTC is gonna be that’s very smart that way all everybody has walked the building we’ve all stood there and said, do we think this is really where badge printing could go? Do we really think this is. Because we also make basically every decision as a team for the conference, so you know there’s no one person on staff whose job is the conference and they get to make the decisions we we do it together or we say who wants to be in this decision, you know I think we’ve even talked about some of the the way Zen10 works um on the show before but. Yeah, so it feels good to have everybody be there, which means this late summer, early fall we’ll all party in Detroit and see what 26 NTC is gonna look like. OK, 26 NDC NTC in Detroit. Uh, no, it’s very smart and you’ve walked out. in the hotels we can make a couple of restaurant suggestions if somebody comes up because we, yeah, yeah, we went to that place. Yeah, very savvy, very savvy. I admire it. I don’t know. I just feel like I would never have thought of that. Like if I was a CEO I would, I wouldn’t need somebody else to suggest that. I think I never would have thought of that. Can I ask you a question? Do I get, is that allowed on nonprofit radio? OK, OK. I had an anarchist in the previous, uh, OK, I was gonna ask you, you know, I know, I mean with 160 plus sessions every year you have a hard job of only doing 30 or so interviews or or 20 or whatever the number is, right? Yeah, yeah, because it’s that’s only a fraction of how many sessions you could have um chosen from. But even still, I’m curious from the interviews you’ve had from, from a day and a half or so. This is content different this year? Like, is there, is there trends or or interesting notes surfacing because of the time we’re in, because of the moment we’re in the moment of under this administration, the the shifts of of the sector, but also even just like this weird moment around AI and the moment around, you know, all of those different pieces, is there something that you’ve seen? Artificial intelligence, yes, more panels on artificial intelligence. I, I requested more so we’ll have a couple more um. Yeah, I, that’s that’s what I was thinking it’s artificial intelligence, um, you know, the, the political environment, it comes out a little bit, but, uh, first of all, a lot of our, well, yeah, no, it comes out you know of course beyond the fundraising, uh, I was gonna say the fundraising panels have mentioned it, you know, but, but, um. Uh, also in the tech, uh, we just had a panel on personally identifiable information, how to preserve that, um, and including from government intrusion and subpoena, um, so you know that would not have been a topic last year would have been a thought, a possibility, um, so yeah, I’d say those mostly the AI and then the the the political and the, um, data, data protection, data protection, yeah, yeah. Um, Is uh is is Max here? Max stage managing? No, he hasn’t for a number of years a few years yes, yes, and I didn’t see Or Louise here. No, she, it’s not her spring break. Oregon already had spring break, so she is very mad to miss her first. TC. Oh, is this the first one? I think so many years at least, or yeah, I think so. She’s about to be 9. That’s right, yeah, yeah, I guess she missed Denver too because that was she was in school then, but yeah, so those are the only two that she’s ever missed, yes, even as a little babe, yeah, so we don’t have like a family photo booth photo from this, yeah, she has them like up on her, on her, you know, cork board in her bedroom. Yeah, I know he has all the NTG. In like the photo, you know, yeah, parent, yes. Um, How’s, uh, you know? How, how are things, uh, how’s the team, how’s the team doing? Thank you for asking. Um. I, I’m just looking because they’re in that room. That’s our, that’s our staff office so that we can see if there’s any issues, but that’s where everybody’s working the halls on fire or something. You’ll see it’s like a uh um called um smoke towers or fire towers. So, so we’re on, we’re we’re down on the floor, but up maybe 40 ft or something there’s windows and that you’re saying that’s the it’s like in the mall, you know, where the security is behind the one-way glass up above, yeah, that’s, yeah, so staffs up there’s too high, we can’t see in you know staff is OK, I think. Like Any any organizations group of staff, you know, there’s. A staff person dealing with this other issue that’s not work related and somebody you know caring for a relative and somebody trying to help their kid that’s having this struggle is that you know so. Uh, the team is OK. We’re a great team and folks have just been like. Even reflecting in real time like God we’re just like showing up for each other so strong and we feel so like happy to get to show up for each other these these few days together but also try to hold that like you know behind that or underneath that or around that. Life is hard for everyone right now in so many ways that have nothing to do with putting on a conference or or replying to your work emails or you know it’s just like everyone is always carrying these these other things that they’re thinking about um and we know and we’ve and we’ve heard from community members as we do every year that. We want the NTC to be a place where like you can set all those burdens down like it’s OK to talk about all of that stuff you know this is not like a perfect professional face you know like what does that mean? What what is my professional? I don’t know. You know, but also I think it’s a little bit hard for staff to feel like they get to do that too when we’re working so hard to create that space for the community because we’re also like on radio and calling catering and you know we’re like doing all these other logistic pieces so in some ways the staff like miss out on the on the best opportunity to do that that we have every year because we we’re we’re kind of behind the scenes but usually on Friday. Less logistics are happening because it’s the final day of the conference and then you know all but one or two people will get to be at the general session and listen to the keynote and everybody will get to like just go to a bird’s table at lunch or go you know people will really get to kind of come out and and. Enjoy it for a little bit. And when does the team depart on Saturday? Yeah, we fly out on Saturday. Oh yeah, we’ll, I mean Friday, Friday afternoon. Oh yeah, our stuff is getting palletized and taken away at 3 p.m. tomorrow. Oh yeah, yeah. Oh yeah, yes, and like we have volunteers, we have, you know, we have um. I mean it’s end 10. We have a spreadsheet, you know, and every storage bin is numbered and we know what goes in bin number 1 and bin number 2 and you know you just look at the spreadsheet and everybody knows what has to go. I see. Of course we have a spreadsheet for that. Yeah, wonderful, um. Yeah, I would say, and I’ve told others this, uh, this is, this is, I think this is the 11th, uh, NTC that nonprofit radio has come to, yeah, uh, I think it’s the best. I do, I do, you’re just saying that because I’m sitting here and I’ll push you off the yoga ball. I said it to somebody else. No, it’s, yeah, you know, you know, uh, no, I really do admire the open plan. I love. It’s, yeah, it’s just a better feeling, you know it’s not a congested feeling we’re not confined to a 10 by 10 cell. Right and like just from an attendee’s perspective, you know, we didn’t want to shrink it down to only a room of 1800 chairs in theater seating or something, you know, like there’s all this open space. Do you wanna just sit on the floor and build Lego? Go for it, you know, you wanna drag a chair away from a table and talk to somebody else? Go for it. Right, we, we, it’s called a pre-con when before you have an event we have a meeting with, you know, the head of every part of the facility, right, uh, like here’s the security, right, like, you know, one rep from every part of the building comes and you have this precon meeting. And they are like, OK, who are you? it’s like just the day before or a couple of days before or is this we do we do it, no, no, no, we do it Monday so you’re on site, you’re like ready to go. You’re starting to load in and usually events that do this like send one rep from N10 right? we send all of us, all 16 of us show up because each of us are here, right? And you’re supposed to say like, you know, is there any info about your attendees? Is there anything the team should know, you know, and we say. If our attendees want a chair move, that chair is moving. Right, you’re not radioing me for like an approval. If attendees walk up to you and say the water is out, you’re refilling the water, right? Like this community is making the space. Our job was just to make sure water stored in there 16 of us right that you can take direction from, yes, and I mean if they say we’d like a whole another part of the building and could you bring catering there that please don’t do without our approval. Right, but we really want to have a, a place that is just open. Make it. Do put the chair where you want it to be, right? Like Ryan, bring your penguins and set them up. Apparently everybody’s like putting in the chat who gets to take them home, you know, like that feels like a gift to ourselves and hopefully a contribution that’s additive to the sector to to have a space that’s like that or that tries to say. What does it look like? We’re, we’re not a trade show. We’ve never been an actual trade show, but what does it look like to say, yeah, there’s no booths, just talk to each other. And honestly like there are some providers who are here for the very first time and they’re like what? What am I doing? What, what am I supposed to do, right? So we need to do a better job of setting people up whether they’re a provider or a sponsor or just, you know, attendee what to expect, what do you, what does it mean to just walk into this huge room and pick, pick a velvet couch to sit on and talk to somebody, you know, but just walk up yeah um. I, I, I just heard someone’s heels. I, I looked because I thought I heard pickle. I thought it was pickleball, but there is pickleball here. We have there’s pickle ball. There’s ping pong. Oh yes, and the pickle ball tournament yesterday, you know, it ended in, I’m not gonna say controversy, um, but it was heated because E0’s very own Carl came in 2nd place and he desperately would like that trophy. Is there a rule about employees? No, no, no, no, but Carl keeps he has spent a year reminding everyone that last year in pickleball he lost in the first round, but to the eventual winner. Right, it’s just, it’s just like, you know, the, the drama of a bracket process and so he wanted to redeem himself this year, right? He made it all the way to section. Yes, right, Carl. Alright, alright, and Carl is celebrating, I think, 18 years as an N0 staff person. He’s our IT director. He started as an AmeriCorps Vista. That’s incredible. Yeah, is he the longest? He’s the longest. And then how many years are you since membership I’ve been, I’ve been, I started at the NTC like the, you know, two days before 11 NTC was my first day as membership directors, um, and, and Ailey’s right behind me at 13. 0, she’s she’s outstanding. She curated this table that we’re sitting at. Uh, she got the chairs that we’re sitting on, um, she chose this spot for you all to be able to see everybody coming in visibility. Yes, it’s very nice. Yeah, Ali on her game, her thank you. All right, I know you’re busy. Thank you, CEO. I hope Aria was a nice surprise. It was such a gift. Thank you for doing that. I so appreciate it. I’m gonna tell the whole team with our at our daily debrief, um, and. And I look forward to the next time Gene and I get to be on the show and I have slept in some reasonable amount of hours prior to talking into the microphone, but I really value you being here and creating a digital platform for so many of our community members to get to share, you know, all of their smarts beyond these walls. So thanks for thanks for the collaboration. Yeah, lots more. OK. Amy Semple Ward, they’re the CEO of N10. They’re our technology contributor here at uh nonprofit Radio. Thank you. Um, and thank you for correcting me, parent, parent, yeah, yeah, you’re right. I mean, I know you’re right, but thank you for the, thank you for the correction. I’m, I’m a trainable boomer. You’re very trainable. We’ll leave it there. Thank you also for being uh with with uh nonprofit Radio’s coverage of 25 NTC where we are sponsored by Heller Consulting technology services for nonprofits. It’s time for a break. Imagine a fundraising partner that not only helps you raise more money, but also supports you in retaining your donors. A partner that helps you raise funds both online and on location so you can grow your impact faster. That’s Donor Box, a comprehensive suite of tools, services and resources that gives fundraisers just like you, a custom solution to tackle your unique challenges. Helping you achieve the growth and sustainability your organization needs. Helping you, help others. Visit donorbox.org to learn more. Now it’s time for the human factors driving your CRM success. Welcome to Tony Martignetti nonprofit radio coverage of 25 NTC, the 2025 nonprofit Technology Conference. We’re at the Baltimore Convention Center, where we are sponsored by Heller Consulting technology services for nonprofits. My guest now is Ruben Singh. He is founder and CEO of 1/10 Consulting. Ruben, welcome back to nonprofit Radio. Thanks for having me, Tony. Good to see you again. This is, I think your 3rd, maybe 4th, 4th you’re counting, you’re more accurate than I am. Um, your session this year is the human factors driving nonprofit CRM success. Uh, I don’t know, I wonder, do you see people often or organizations often blaming technology when the problem is more team and human? Yeah, that’s exactly the the premise of it, Tony. I, I, um, uh, I’ve I’ve often found myself as a consultant coming into situations where Um, you know, the client we’re working with at the prospect we’re speaking with says, ah, we need Salesforce or we need, you know, uh, this particular solution or virtuous or something else, um, and, uh, you know, because they’ve had a bad experience with the technology that they were with, uh, and so, uh, as I’ve seen this so much throughout my career, um, and then you know you start peeling back the layers of the onions. And then you see, oh well, you know, the business processes are not fully defined. There’s not, you know, a plan for adoption, there’s not a plan for governance, uh, and so what I’ve realized is that, um, often times, uh, where the, the, the failures happen have really has nothing to do with the technology at all. So yes, that’s exactly what the premises of this uh of this um workshop. So the symptom I don’t know, cycling through, uh, platforms like, oh we need Salesforce, and then they have Salesforce for 3 or 4 years and then they realize, oh, no, Salesforce wasn’t really the solution. Now we need we need the Microsoft platform, right? Does it happen like that? That’s definitely a part of it. I think there’s also, you know, just there’s always this sense of urgency um as new technology comes along. Um, that uh we don’t, if we don’t innovate, we’re gonna fall behind, um, and we’re seeing a lot with, you know, AI now as well. Um, it’s, it’s no different, you know, everyone say well if you’re not using AI for your solutions, uh, you know, you’re, you’re gonna be left behind, you’re gonna, your organization’s gonna fail when really. Um, that is, that is really forcing organizations to adopt certain things that they’re not, they may not be ready for their, their processes may not be in place, their data might not be ready. So, um, so just like we’re seeing right now with, you know, the latest trend of technology, um, feel the organizations we work with just feel compelled to to to rush into whatever the particular trend is, um, and sometimes there’s business pressure also. Uh, you know, if this particular system failed, you know, new, new CTO comes in and says, ah, I need to prove myself or I need to get something going, so oftentimes they try to get the, the, the cart before the horse. You mentioned business processes a couple of times. What what kind of processes should we, so now we’re getting to what the human factors are that you need to have in place for the technology to be successful for the CRM. Well, what kinds of business processes are you looking for? Sure. Well, um, first, are they defined, are they documented? You’d be surprised how many nonprofits I walk into where, you know, they could be operating just fine, um, but there’s no nothing really documented and that’s fine, that’s where, you know, consultants like ourselves who would come in and help, uh, understand them, review them, define them. Um, and then there’s some that are just, uh, not very, uh, as you as you learn more about their business processes, you see that there’s, um, just inefficiencies, uh, that are there as well. So what kinds of business processes are we talking about? Just some examples. Yeah, um, fundraising, for example, so check come, you know, stack of checks come in, you know, what are the different processes to get that data entered into the system, or it might be donor advised funds, you know, what is the process around that? Or you know, we have a uh uh a series of files that need to be uploaded or it could be an application. process for programs uh that that might be so it it it could be um the grant distribution process what’s the reviews and applications that process who needs to review, who needs to approve before a grant might be approved. So, um, so there’s really a cross grant making fundraising program delivery um there’s there’s all kinds of steps that needed to happen and we ultimately want the the system, the technology to make it work as efficiently as possible. So we we we heavily rely on business processes to make sure that we’re. Um, that we’re creating things that ultimately makes the systems more efficient. So the technology is supporting the business processes not not we’re not relying on the technology for the business process, but it’s, it’s supplementing what our processes are. That’s exactly it. I would, I would add one more thing to that, you know, in addition to the technology supporting the business process, it also has to support the strategy. So that’s kind of one of the other points that we’re gonna be talking about later on today. Um, is to make sure that there’s a clear strategy. What is it that as an organization you’re trying to solve, um, how do you know that you’re successful? What are those markers or indicators to confirm that you’ve, that you’ve, you’re successful because we want again the technology to support that, uh, we want the technology to be able to track and report and monitor to make sure you’re meeting your goals, so. Uh, as a technologist, it’s a little bit tricky sometimes because they, they, you know, the clients often expect us to come in with a technical solution and, you know, code this and configure that, but we, we like to kind of step back and say, you know, what is it that you’re trying to solve? What’s your strategy? How do things work? Where can we create some efficiencies, then we start building. OK. Yeah, it’s it’s, it’s often important to have that consultant perspective to take a step back because you’re, you’re probably often. Uh, at 110th Consulting, and I’m, I’m gonna ask you to remind listeners why it’s 110. I think I remember, but we’ll get to that, we’ll get to that shortly. Um, you know, you’re, you’re often, um, confronted with, you know, we need, we need new technology. That is not our, that’s not our feedback. Amy. It is, it is feedback. OK. OK. There we go. Thank you. Oh, can you just try it with a lower volume? We’re uh we’re we’re, you can tell we’re live here at we’re, we’re just turning on the loudspeaker because the plenary session just ended, so we have a we have a loudspeaker and it was feeding back. OK, so that was us, that was us. Tala. All right, we’re accountable. I’m accountable. I feedback. It’s time for Tony’s Take too. Thank you, Kate. We’ve got a new tale from the gym introducing you to Roy. Uh, I met Roy by. Same way I hear about lots of people, uh, overhearing other people talking about Roy, uh, after he had left. And I know who he is. I’ve, I’ve seen him around. Uh, I’d say Roy is probably early to mid 70s. And the, the thing that the uh the two guys talking about Roy were focused on. Was that, um, well, he talks a lot, I agree. Uh, and he doesn’t put the weights away after he uses this one machine that he focuses on a lot, spends a lot of time on, and, and he actually grunts a lot when he’s on this bench press type machine. Uh, I had not noticed that he doesn’t take the weights off like you’re supposed to, but these two, guys did, uh, as well as talking about how chatty he is, and, you know, that he, he like, uh, just spends too much time in between his sets. Talking to other people. Maybe even annoying other people, uh, they did, they didn’t say that, but. They didn’t like how much he talks, and they don’t like that he doesn’t put the weights away. So I was, uh, so I was keeping an eye on Roy, you know, after I saw, I saw him again, and, um, word must have got to him about the weight part. Uh, he didn’t, he didn’t talk any less, still very chatty between sets. But he did start putting his weights away after he was done, takes them off the machine, puts them back on the rack where they get stored. So that’s good. So Roy did uh improve his bad gym behavior. Now, how did he get wind of the idea? That he’s not. Practicing good gym etiquette, I don’t know. Uh, I, I, I had nothing to do with it. I just stay out and listen and, and I happened to watch well, cause I heard about Roy once I learned his name, so then I was paying more attention, and that’s when I saw that he, he, uh, exercised good gym etiquette. So, but how he came to change his ways, I don’t know. I swear I had nothing to do with it. So that’s Roy at the gym, uh, along with many of our other characters I’ve introduced you to, uh, through the, through the many months. Roy at the gym, the community gym where I work out, uh, 44 times a week. And that is Tony’s take 2. OK. Are you sure you haven’t spoken to Roy before? Because he sounds familiar. I don’t think so. You might be, you might be thinking of Rob, Rob. There was Rob. He was the Marine Corps. Remember Semper Fi? He was talking to a another former Marine, and they ended their conversation with Semper Fi, the the Marine Corps motto. You might be thinking of Rob. I, I’ve seen Roy, but I never knew his name. OK. Hm. A lot of our names that are short. Well, it’s a couple. Try to keep track of all the characters at the community gym where I go, try to, maybe we should start a database. We should do, maybe we should start a CRM database. We said. We’ve got Fuku butlers more time. Here’s the rest of the human factors driving your CRM success with Rubin Singh. So you’re often confronted with our technology sucks, you know, we need this platform is not working for us. Nobody, I don’t know, nobody ever seemed to have learned it right. It doesn’t, it doesn’t work with what we’re doing, what we’re trying to do, and but you need to step back and say that there may very well be something deeper than your platform. OK. Absolutely, and you know there there’s there’s that and and I’d say also there’s been times if I’m being totally honest, Tony, where I’ve don’t don’t don’t be disingenuous, don’t lie to no I can’t do that, um, but, uh, you know, I’ve I’ve been part of teams early, early in my career where, um, you know, we have, we’ve completed the project. We have finished on time, we’ve finished on budget, we checked all the boxes for all the requirements that we met, but at the end of the project, at the end of the go live, honestly I, I felt kind of uncomfortable. I felt uneasy. I felt sick to my stomach because I knew that despite all the things that we’ve done for the technology, the nonprofit is not set up for success in the long run. Uh, some of the ways that I sense that were, um, they, you know, maybe all the users were really excited about the technology, but maybe the leadership was not fully on board, so they may not have, um, you know, they have not modified their processes, how they’re gonna do reporting, how they’re going to measure progress, how they’re going to measure performance, um, if they’re not really bought into the system, um, the whole thing’s going to fail. Um, also another key thing is governance. For leadership, the leadership buy-in is essential, uh, I think we may have talked about, or if it wasn’t you, it was another NTC how to get that leadership. It’s, if, if, if the leadership isn’t committed, leaders who are listening, CEOs, uh, executive directors listening, if you’re not completely committed, I mean your teams, your teams know that they figure it out and their commitment, uh, uh, is gonna be equivalent to yours for sure. If you could spend tens of $1000 on a on a new CRM system, if not more, but if you as a leader are gonna say, OK, you know, fundraising meeting, you know, fundraising, you know, check-in is going to be on Monday and everybody bring your spreadsheets, um, forget, you can forget the investment that you made in that CRM system, you’ve lost adoption right there holistic look remind us why your your company is one. Yeah, you know, well, um, you know, we’ve exclusively worked with nonprofits and, and very early in our, in our startup phase, uh, we worked with a lot of, um, uh, faith-based organizations, um, and, uh, it was, it was funny that as I was working with different organizations whether it was a synagogue or a church or a mosque or or cordura. Um, they all seem to have this 10% concept, this giving back of 10%, uh, to back to the community, back to causes, back to the good of, of the whole, um, and I thought that, you know, despite all these different faith traditions being so different, there was something that was, uh, you was a common thread amongst many of them, and so that’s kind of where 1/10 the consulting came from, yeah. I had it close. I thought it was, I thought it was giving I was giving back, but giving back 1/10. Yes, exactly, exactly, um, you, you, uh, your session description talks about inclusive design as a as a means to. Achieve the the CRM success that we’re looking for. Uh, say, say more about the design process, inclusivity. Yeah, well, well, we feel strongly that any technology that is meant for everyone should uh include as many people as possible as part of the design, otherwise, um, there’s gonna be blind spots. Um, uh, you know, this might seem, uh, you know, unrelated, but I was reading some stories about how when Um, uh, there’s crash test dummies, um, were being used, um, to test the safety of vehicles. They were based on a male body of 5′ 970 pounds and, um, the, and as is the 2011 University of Virginia study that showed that women were much more likely to be harmed or hurt in an accident. Um, than men were and 47% more it was like a pretty ridiculous number, so, um, just makes you wonder like when, when these vehicles and safety were designed and assessed, were they really thinking about different body types, different people? Well, the same thing applies to, you know, technology, um, the story I often give people is, um, you know, even my own parents, my, my elderly, uh, you know, parents, you know, from an immigrant community when they signed up for the COVID vaccine. Um, it was a process that was clearly not designed for immigrants and it was not designed for elderly, um, you know, it was, it was a the application form was very, uh, cumbersome. There was a lot of information you had to have, have prepared you cited this in a year or two in the past, yeah, the online form was not, uh, well, not user friendly for 70 or 80 year old, yeah, yeah, and so. Um, and, and so like, and then you, you also hear that uh well the immigrant communities are not signing up for the vaccine and, and it’s a public health problem. So, um, you know we feel that CRM is the same day is the same way that um it often times I walk into a room for a design session and I have the IT professionals there or because they’re the ones who can give us the quickest answers, um, or we have the people who are like the quote unquote super users who are the most technically proficient. Um, but you know, to me, if we really want to have inclusive design, we really want to see what who who’s not at the table. Let’s maybe have people of different ages, different technical proficiency, different, you know, socioeconomic, different uh parts on the company hierarchy because if this system is really designed for everybody, we need to get as many different thoughts, ideas, perspectives involved, um, to me, uh, often times if we overlook that, um, it, it, it ends up being a gap that we have to fix later. Um, you, uh, you also cite, um, this being valuable for uh professional growth. How is that? Yeah, and this is something, uh, as I get older, Tony, I’ve been reflecting on on a lot and uh I’ve been, um, I’ve I’ve felt this in my professional career and I’ve I’ve had the um privilege to uh teach as an adjunct professor close by here at University of Maryland Baltimore County. And it’s something I often tell my students, it’s like, um, you know, if, if you really want to be good consultants in this space, um. You know, we, we often times focus on the, the, the credentials or the certifications and you know we go crazy with the certifications and, and you know so we can present that and say, ah this is, you know, this is who we are but in my experience what’s been most useful and also has been useful to to the success of our projects is really being an expert in the industry that you’re working with, being passionate about the industries that you’re working in. Um, so for for us we tend to have a focus on the social justice sector, um, and it’s, you know, in my tradition, in my blood, and my, in my, uh, upbringing to be working with these types of organizations and really understanding the ins and outs of, of these organizations, and I feel like that’s been able to, um, uh, it’s it’s been something I’ve been uh able to bring to my projects, bring to my implementations, um, and the love and care that we give our clients and. It’s, it’s been very helpful so it’s something I, I encourage folks to, um, you know, to, to work on becoming experts work on becoming um keeping up to date with what’s going on in the industry so if you’re a nonprofit technology consultant being fully aware with the challenges with grant making or you know how government funding might be affected these days or you know taking an opinion on how AI can help or harm uh nonprofits. Uh, taking an opinion on data privacy and where it fits, so, uh, what I’ve seen is that what our clients need is not just, just people who are experts at the technology, but experts at the industry that they’re working in. And what about professional growth for for folks in nonprofits as they’re looking at their own businesses? I mean this is sort of a broadening exercise where folks are learning. there as you’re you’re suggesting for consultants, the folks nonprofits learning outside their own areas of expertise. Yes, yes, absolutely, um, an example of that is, you know, uh, to me I realize that, you know, equity and and technology is, is, is ultimately, you know, uh, it’s super important for inclusive design and inclusive systems, um, so for me that meant, um, I was gonna become a. A certified DEI practitioner, um, so I went and, you know, went through the classes, got my certification, and, and that was something that I, I wanted to make sure that was part of who, who I was. So while you might think that this might not be part of your part of your, um, technical credential, um, this might not be part of your technical credential, um, it, it having that business credential or having that expertise was super important, um. Another example could be change management. Um, so if, if you feel that change management, excuse me, change management is something that you’re passionate about, by all means become a certified credential change management professional because it’s only going to make all your projects, uh, more successful. Get you outside your comfort zone. Yeah, I mean I think there’s I think there’s value in that. I’ve I’ve seen it professionally, um. Just, you know, challenging, challenging yourself, you know, outside outside your normal boundaries and and and it really applies to any industry, so as I tell my students, if you’re a The technology is everywhere, so you know if you wanna do fashion tech, become an expert in the fashion side of it a sports tech, become an expert in the sports side of it doesn’t really matter, um, but it’s really just about, you know, what my clients over 27 years of of of this work, um, they don’t really care how many salesforce certifications I have. They don’t really care about, you know, what they, what they care about is do I understand them? Do I understand their business, do I, you know, what, what else do I bring to the table? There’s lots of tech consultants out there, but what else do you bring to the table? So it sounds like we’re in your backyard, uh NTC. You teach at University of Maryland Baltimore County. Yes, yes, yes, just uh uh right right outside 20 minutes from here, so it’s nice, we’ll see you again next year. I don’t do you know where next year’s is? I believe Detroit, but, uh, wherever it’s gonna be, I’ll be there. We’ll be together. Um, what else, what else are you gonna share on this topic that, uh, you and I haven’t talked about yet? Um, we are gonna, I think the one other thing, and, um, it’s not something I talked about, you know, a few years ago, but it’s just so much more relevant now is, um, you know, again as far as the, the human factors that that uh affect CRM, it’s also uh looking at um, you know, how bias and discrimination can make their way into the systems, um, so whether it is algorithms that are built, whether it’s AI models, um, and, uh, making sure that we have. Checks and balances in place to ensure that the data is um not toxic in any ways or or the data is not skewing results in a in a way that could um hurt or harm communities. I think one example I if I can give a specific is, you know, if you are a nonprofit that uses an algorithm for recruitment volunteer recruitment or application reviews, um, you know, making sure that those do not, uh, uh, create skewed results, making sure that. There’s a checks and balances process to make sure that that that the results are not discriminatory, that they’re fair, they’re unbiased, um, and that’s something that, um, organizations are really grappling with how do we do that? So I have some, some models that I’m gonna be sharing as well today that I’m pretty excited about on on how you can create some checks and balances. Yeah, um, I mean it’s uh I wish I wish I had my diagram I could bring up here too. Yeah, but um, yeah, yeah, um, but it’s basically just um being intentional about, OK, you know, if again starting with strategy, so let’s say it’s a volunteer recruitment plan and you say, you know, I and I had a lot of this well we want to increase our, our diversity in our recruit in our volunteer pool so you know it’s, it’s basically setting some measures for that. um, we wanna have 20% of, of this particular demographic or 40% of this and then it’s really just um creating some checks and balances just like in any. Um, in any implementation of of technology there’s gonna be a, a testing phase, there’s gonna be a data validation phase. So what what I’m proposing is we also have a a a a a bias detection phase. We also have a, um, you know, um, uh, a sort of um discrimination and bias, uh, check that we essentially do so in that case we will just like we have parameters to say these are the percentages we’re looking for, what are the results for it? So, um, so it’s, it’s not rocket science, but, but it’s really just making sure that when we build our test plans out that we’re also checking for bias and discrimination. I know a lot of work on LinkedIn. Have you written a book? Um, in the works, Tony, in the works you mentioned it or maybe I’m just that you ought to write a book. OK, uh, you’re working on a book, um, early stages, yes. But um but yeah, now there’s I’ve I’ve kind of captured a lot of thoughts um over the years about this and you know these are things that were just kind of things I uh thought about, you know, like uh you know this doesn’t feel right or this could be done in a more equitable way or this is this is not really doing good. Even though we’re calling it tech for good, um, and then, you know, being in communities like this at N10 and NTC, uh, made me realize, oh, I’m not the only one who thinks like this. There’s others who who are who have also found some weird stuff in out there in the implementation world and and I think, um, you know, as a consultant who, who now has started my own practice, um, I realized, you know, I don’t have to just continue being part of the problem, you know, I can. Uh, I can try to shift some things and, and, um, share my stories to, to make sure that we, we, we, we collectively do better. OK, now, you know, I imagine you’re, you’re part of a minority faith community in the United States. How does that inform your practice or or how does it open your eyes to the inequities that that we were you know, you’re doing more to fight than I am. I bring voice to them, but you’re actually doing. Well, well bringing to it is is very important, so I appreciate that you do that, Tony. Um, uh, yes, as part of the sick tradition I think um it’s it’s a few things. um, I mean it really does fuel a lot of my work, um, you know, the sick tradition is a very um deep in in community service and in justice and in collective liberation, so it it really is a fuel for, you know, the, the, the focus on the social justice sector. Um, but I would say, yeah, absolutely, you know, being a visible minority does, uh, help me have, uh, that the radar is always on. Um, you know, and, um, I, I’m, I’m very aware, hyper aware of of things that just don’t feel right or that that that that don’t sit well with me, whether it’s happening to me or someone else, um, so you know, early on in my career I was, you know, thinking to myself, ah, you know, I’m just learning the ropes, let me just follow along what everybody else is doing and you know, different practices that might happen in in implementations themselves and technology implementations. I’ll just go along with it. It’s fine, um, but now I, I didn’t it right, yeah, so now I kind of sense that, you know, I feel that agency that I can, you know, I can, I can speak my mind, I can step forward and say, yeah, you know, this, this persona building exercise we’re doing for marketing, you know, of, you know, guessing what. Different races and demographics might feel about our work. Yeah, it doesn’t, that’s not great. That’s there’s other ways to to get that information that doesn’t sit here and, and, you know, enable stereotypes, uh, you know, we can use archetypes, we can, you know, we can ask people directly why they could come to our nonprofit or don’t come. So there’s there’s alternatives out there, you know, the tried and true methods are not always the best, um, they’re not always the most equitable, so you know, let’s let’s brainstorm other alternatives. Did finding that agency come from starting your own business or before then? I think it was a combination of being in circles of other technologists of color that, you know, where I, I felt very empowered and said, ah, you know, we’re not the only ones, you know, like other people feel this way and collectively. You know, there’s things we can do better, um, and then yes, starting the company and I I recognize that comes with privilege, um, and not everybody can speak out the way that they want to, but, but definitely starting starting my own practice and being very transparent with my customers about, hey, this is who we are, this is what we’re about. If we see something that doesn’t feel right or that the data that you’re requesting from your clients is is overreaching, we’re gonna, we’re gonna raise our hand and and you know what I thought might have deterred. Customers is actually um uh had customers gravitate towards us. They want to be held accountable. Yeah, yeah. Well, I’m glad you found your voice it’s always good to see you, thank you very much. Thanks so much. My pleasure. Singer and CEO. Thank you for joining us for our. 2025 nonprofit technology conference coverage in the Baltimore Convention Center and thanks to Heller Consulting technology services for nonprofits for sponsoring nonprofit radio at 25 NTC. Next week, more from 25 NTC PII in the age of AI. If you missed any part of this week’s show, I beseech you. Find it at Tony Martignetti.com. We’re sponsored by DonorBox. Outdated donation forms blocking your supporters’ generosity. Donor box, fast, flexible, and friendly fundraising forms for your nonprofit, Donorbox.org. I’m gonna miss that alliteration. Donor box is going away this week. Fast, flexible, friendly, fundraising forms. Our creative producer is Claire Meyerhoff. I’m your associate producer Kate Martignetti. The show’s social media is by Susan Chavez. Mark Silverman is our web guy, and this music is by Scott Stein. Thank you for that affirmation, Scotty. Be with us next week for nonprofit Radio, big nonprofit ideas for the other 95%. Go out and be great.

Nonprofit Radio for September 30, 2024: AI, Organizational & Personal

 

Amy Sample WardAI, Organizational & Personal

Artificial Intelligence is ubiquitous, so here’s another conversation about its impacts on the nonprofit and human levels. Amy Sample Ward, the big picture thinker, the adult in the room, contrasts with our host’s diatribe about AI sucking the humanity out of nonprofit professionals and all unwary users. Amy is our technology contributor and the CEO of NTEN. They have free AI resources.

 

Listen to the podcast

Get Nonprofit Radio insider alerts

I love our sponsor!

Donorbox: Powerful fundraising features made refreshingly easy.

Apple Podcast button

 

 

 

We’re the #1 Podcast for Nonprofits, With 13,000+ Weekly Listeners

Board relations. Fundraising. Volunteer management. Prospect research. Legal compliance. Accounting. Finance. Investments. Donor relations. Public relations. Marketing. Technology. Social media.

Every nonprofit struggles with these issues. Big nonprofits hire experts. The other 95% listen to Tony Martignetti Nonprofit Radio. Trusted experts and leading thinkers join me each week to tackle the tough issues. If you have big dreams but a small budget, you have a home at Tony Martignetti Nonprofit Radio.
View Full Transcript

And welcome to Tony Martignetti nonprofit radio. Big nonprofit ideas for the other 95%. I’m your aptly named host and the pod father of your favorite abdominal podcast. Oh, I’m glad you’re with us. I’d suffer with a pseudoaneurysm if you made a hole in my heart with the idea that you missed this week’s show. Here’s our associate producer, Kate to introduce it. Hey, Tony, this week A I organizational and personal artificial intelligence is ubiquitous. So here’s another conversation about its impacts on the nonprofit and human levels. Amy Sample Ward, the big picture thinker, the adult in the room contrasts with our hosts, Diatribe about A I sucking the humanity out of nonprofit professionals and all unwary users. Amy is our technology contributor and the CEO of N 10 on Tony’s take two tales from the gym. The sign says clean, the equipment were sponsored by donor box, outdated donation forms, blocking your supporters, generosity, donor box, fast, flexible and friendly fundraising forms for your nonprofit donor box.org here is A I organizational and personal is Amy sample ward. They need no introduction but they deserve an introduction. Nonetheless, they’re our technology contributor and CEO of N 10. They were awarded a 2023 Bosch Foundation fellowship and their most recent co-authored book is the tech that comes next about equity and inclusiveness in technology development. You’ll find them at Amy Sample ward.org and at Amy RS Ward. It’s good to see you, Amy Ward. I do love the Pod Father. I know it makes me laugh every time because it just feels like, I don’t know, like I’m gonna turn on the TV and there’s gonna be like a new, new, new season of the Pod Father where we secretly, you know, follow Tony Martignetti around or something. We are in season 14. Right? Yeah. Um Yes, I appreciate that. You love that. It’s, you know, you like that fun. So uh before we talk about um the part of your role, which is the, the technology contributor to N 10, uh the technology, the nonprofit radio and we’re gonna talk about artificial intelligence again. Let’s talk about the part of your life that is the CEO of N 10 because you have uh have you submitted this major groundbreaking transformative funding federal grant application? Yes, we submitted it last night three hours before the deadline, which was notable because I, I know there were people down to the, the minute press and submit. No, we got it in three hours early to what agency um to NTI A they had, this is kind of all the work that rippled from the digital Equity Act that was passed in Congress a couple of years ago. And, you know, now, you know, better than to be in Jargon jail. What is NTI A, it sounds like an obscure agency of our, of our federal government. It’s not, well, maybe to some listeners it’s obscure but it is, um, the National Telecommunications and Information Administration. And you, I think that’s obscure to about 98.5% of the population, you know, I think I, I think I’m obscure to, you know, uh being obscure is fine. Um Yes, the National Information Administration and um prior to this uh grant um from the federal level where folks from all over were applying, every state was also creating state equity plan, digital equity plans. Um What funds might be available through the state funding mechanism to support digital equity goals. But a lot of those at the state level are focused on infrastructure, like actually building internet networks to reach communities that don’t have broadband yet, you know, things like this and so very worthwhile funding endeavor. I mean, we need, we need to have 100% of the population needs but even with those state plans and the work that will come from them and the funding it will not, we are not about to have every person in the country have broadband available to where they live, right? Ee even with all of this investment, it, it’s not gonna reach everyone and that means that the amount of funding within state plans for the surrounding digital literacy work, digital inclusion work, you know, making sure people know how to use the internet, why they would use it have devices. All those other components is gonna be really minimal through the state funding because even if they used all of it on infrastructure, they wouldn’t be done with that, right. So um the federal government, yeah. So, so the kind of next layer in all of that is this federal pool where they’re anticipating grant making about 100 and 50 grants somewhere averaging between five and, and 12 million each. There’s gonna be exceptions, of course, there’s ma there’s big cities, there’s big states, you know. Um but though all those grants will be operational from 2025 through 2028. So four kind of concerted years of, of national Programmatic investment. Um And these are projects kind of on the flip side, those state projects where this isn’t necessarily about infrastructure and, and building networks or even devices very much, right? It’s mostly the infrastructure programming and you’re asking for a lot of money. So tell, you know, share the, share the numbers, what you’re looking for, how much money. Yeah, we’re our project in the end I think came out at about $8.2 million project and we’re hopeful, of course. Um and I’m, I’m truly curious, um listeners who are always tuning into nonprofit radio from like fundraising strategy perspective. I’d love to learn from you or, you know, email me at Amy at N 10 anytime I’d love to hear your thoughts when you listen to this. But you know, N 10 is a capacity building organization is we, we don’t apply for grants often because quote unquote, capacity building is not considered a, a programmatic investment to most funders, right? And so it’s just not something that um they will entertain an application from us on. And but with this, we have already run for 10 years of digital inclusion fellowship program that is focused on building up the capacity of staff who already work in nonprofits who are already trusted and accessed by communities most impacted by digital divides to integrate digital literacy programming within their mission. Are they a housing organization? Are they workforce development? Are they adult literacy, you know, refugee services, whatever it is, if you’re already serving these communities who are impacted by digital divides and you’re trusted to deliver programs, well, you don’t need to go have a mission that’s now digital equity. No, you digital equity can be integrated into your programs and services to, to reach those folks. Um And so we’ve successfully run this program for 10 years and had um you know, over 100 fellows from 22 different locations around the US and have seen how transformative it’s been. These programs have been sustained for all these years by these organizations, they now see themselves as like the leaders of the digital equity coalitions in their communities. They, you know, fellows have gone on to work in digital equity offices or, you know, organizations et cetera. So it feels great, you have tons of outcomes from a smaller scale program and the grant is to scale up, scale this thing up. Yeah. Yeah. So instead of, you know, between 20 to 25 fellows per year with this grant, we would have over 100 a year. Um And that also means that instead of, you know, if there’s only 20 fellows and maybe we can only cover 20 locations while with over 100 we can cover or at least give opportunities to organizations in every state and territory to, to be part of this kind of capacity building opportunity. All right, it sounds, it’s, it’s huge. It’s, it’s, it’s really a lot of money for N 10. Um uh It, it falls within the range, I guess a little, no, it’s like right within the middle of the range, you cited like 5 million to 12 million, you said? So, yeah, exactly. So our, our application is kind of in the middle there. Yeah, slightly to the low side of middle. But, you know, we just call it middle, between friends. Um Yes and I mean, we’re hopeful, knock on wood, we’re really hopeful that this is an easy application to approve because we’re not creating something new we’re not spending half of the grant in planning. We know how to run this program. We’ve refined it for 10 years. We know it’s very cost efficient, you know, and in the end of four years, 400 plus organizations now running programs that can be sustained is accelerating towards, you know, addressing digital divides um versus, you know, a small project that just end 10 runs. All right, listeners, contact your NTI A representative, the elected person at the National Telecommunications and Infra Information Information Agency. Yes, speak to your uh Yeah. Yeah, let’s get this. Let this go. All right. When do you find out when? Well, you know, there was very clear information about down to the minute when applications were due, but there’s not a ton of clarity on when we will find out. So, you know, they are, they are meant to programs that are funded are, are meant to get started in January. So I anticipate we’ll hear, you know, in a couple of months, of course, and I will let you know, we’ll do an update. I’ll let you know you have my personal good wishes and I know nonprofit radio listeners wish and then good luck. Thank you. I appreciate all the good vibes would reverberate through the universe would be a transformative grant in terms of dollar amount and expansion of the program. Transformative. Yeah, 100% and staff are just so excited and hopeful about what it could mean for just helping that many more organizations, you know, do this good work. So we’re really excited and I admire intend for reaching for the sky because you have like a 2 to $2.5 million budget, annual annual budget somewhere in there. Um And you’re reaching for the sky and great ambitions uh only come to fruition through hard work and uh and thinking big. So thank you, even if you’re not, I don’t even want to say the words if you know, they should blunder if NTI A should blunder badly. Uh I still admire the, the ambition. Thank you. And no matter what, it’s a program that we know is transformative for communities and we wouldn’t stop it even if you know, they make a blunder and don’t, yeah, don’t tell. All right. Listen, don’t tell your NTI A representative. You said, don’t share that part of the conversation. All right. Thank you for sharing all that. And thanks for your support. It’s time for a break. Imagine a fundraising partner that not only helps you raise more money but also supports you in retaining your donors, a partner that helps you raise funds, both online and on location, so you can grow your impact faster. That’s donor box, a comprehensive suite of tools, services and resources that gives fundraisers just like you a custom solution to tackle your unique challenges, helping you achieve the growth and sustainability, your organization needs, helping you help others visit donor box.org to learn more. Now, back to A I organizational and personal. Let’s talk about artificial intelligence because this is not anybody’s mind. I can’t get away from it. I cannot. Uh I’m not myself of the concerns that I have. Uh They’re deepening my good friend George Weiner, uh you know, has a lot of posts, uh the CEO at the whale who I know you are, you are friendly with George as well. Talks about it a lot on linkedin uh reminds me how concerned I am uh about, you know, just the evolution. Uh I mean, it’s inevitable. This, this thing is just incrementally. This thing. This technology is uh is incrementally moving, not slowly but incrementally. I I and I, I cannot overcome my, my concerns and I know you have some concerns but you also balance that with the potential of the technology, transformative techno, the the transformative potential there. I’ll throw you. I was just gonna say, I totally agree. This is unavoidable. I can’t, you know, I cannot go a day without community organizations reaching out or asking questions or whatever and a place of reflection or, or a conversation that I’ve been having and I, I wanted to offer here, maybe we could talk about it for a minute. So, so listeners benefit by kind of being in, in one of these sides with us in the conversation is to think about the privilege of certain organizations to opt in or opt out of A I in the same way that we had for many years, you know, talked about the privilege of organizations in or, or not with social media generally. Like we think about Facebook and we go back, you know, 10 years, there were a lot of organizations who felt like they didn’t have the budget and like, practically speaking and they didn’t have the staff, well, certainly not the staff time but also not the staff confidence. Um I don’t even wanna say skills, but like even just the confidence to say, I’m gonna go build us a great website. They had a website, like they had a domain and content loaded when you went to it, right? But it wasn’t engaging and flashy and interesting and probably updated once, you know, and then Facebook was like, hey, you could have a page and oh, you can have a donate button and, oh, you can have this and oh, and you can post videos and you can, you know, it was like, well, why wouldn’t we do this? Right? And a bunch of our community members spend time on Facebook or maybe don’t even look for information on the broader web, but look for things within Facebook, you know, and, and have it on their phone and are using an app instead of doing an internet search, right? Like they’re, they’re going into Facebook and searching things. So they didn’t, those organizations didn’t feel like they had the privilege to opt out of that space, they had to use it because it came with some robust tools that did benefit them at the cost of their community data, all of their organizational content and data, right? Like it, it had a material cost that they maybe didn’t even understand. Right? And, and didn’t fully negotiate as like terms of this agreement. We’re just like, well, we have a donate button on Facebook and we don’t have one on our website, right? Not, not only, not only didn’t understand the terms, didn’t, didn’t know what the terms were right? Early days of Facebook, we didn’t know how and how many times how pervasive the data, data collection was, how it was going to be, how it was gonna be monetized, how we as the individuals were gonna become the product. And how many times did we talk? You know, I’m saying we like N 10 or, or folks who are providing kind of technical capacity building resources say you don’t know what could happen tomorrow, you could log in tomorrow and your page could look totally different, your page could work different, your features could be turned off. Facebook could just say pages don’t have donate buttons. And you know, I think folks felt like that was very, you know, oh, you’re being so sensational and then of course they would wake up one day and there wasn’t a button or the button really did work different, right? Like you people realize we’re not in control of even our own content, our own data. That’s right. The rules change and there’s no accountability to saying, hey, we need, do you want these rules to change? No, no, no, no, no. Like they set the rules and that was always of course a challenge. But we’re in a similar place with A I where folks aren’t understanding that the there’s, there’s no negotiation of terms happening right now. Folks are just like, oh, but I, I don’t have the time and if I use this tool, it lets me go faster. Because what do I have a, a burden of of time, I have so much work to try and do and maybe these tools will help me. And I’m not gonna say maybe they won’t help you. But I’m saying there’s a incredible amount of harm just like when folks didn’t realize, oh, we’re a, you know, we provide pro bono legal services and we’re based on the Texas border. Now, every person who follows our page, every person who’s RSVP do a Facebook event. Like all these people have a data trail we created that said they may be people that need legal services at a border, right? The there’s this level of harm that folks that are hoping to use these tools to help with their day to day work may not understand. I do not understand. Right. That’s coming in in silent negotiation of, of using these products. Right. And I think that’s, well, I can’t just in 30 seconds say, and here’s the harm like it’s, it’s exponential and broad because it could also the, the product could change tomorrow. Right. It’s this, it’s this vulnerability that isn’t going to be resolved necessarily. You, you said the word exponential and I was thinking of the word existential. Yeah. Both because I think I’m, I have my concerns around the human. Yes. Trade off is a polite way of saying it. Uh Surrender is probably more, is more in line with what I’m what I feel. Surrender of our humanity, our, our, our creativity, our thinking. Now our conversations with each other. One of the, one of the things that George posted about was a I that creates conversations between two people based on the, the, the large language that, you know, the, the, the data that you give it. It’ll have a conversation with itself. But purportedly, it’s two different people purportedly. Uh and I’m using the word people in quotes, you know, it’s a, a, a conversa. So the things that make us human. Yeah, music, music, composition, conversation, thought, staring and, and our listeners have heard me use this example before, but I’m sticking with it because it’s, it, it still rings real staring at a blank screen and composing, thinking first and then composing. Starting to type or if you’re old fashioned, you might start to pick up a pen, but you’re outlining either explicitly or in your mind, you’re thinking about big points and maybe some sub points and then you begin either typing or writing that creative process. We’re surrendering to the technology, music composition. I don’t compose music. So I don’t know the, but it’s not that much similar in terms of creative thought and, and synapses firing the brain working together, building neural nodes as you exercise the brain, music composition is that that probably not that much different than written composition. Yeah, brain physiologists may disagree with me but I think at our level, we you understand where I’m coming from and I’m kind of dumping a bunch of stuff but you know, but that’s OK. II I am here as a vessel for your A I complaints. I will, I will witness them. We can talk about them artificial intelligence. Also from George, a post on linkedin that reflects on its own capacity that justifies you. You ask the um the tool to reflect on its own last response. How did it perform? You’re asking the tool to justify itself to an audience to which it wants to be justifiable in, right? The tool is not going to dissuade you from using it by being honest about it, how it evaluates its last response. Well, yeah, I mean, I think, I don’t know, generative A I tools, these major tools that folks you know, maybe have played with, maybe use whatever you know, are programmed, are inherently designed to appease the user. They are not programmed, to be honest, they are. That, that’s an important thing to understand my point. We have asked the tool, what’s two plus two? Oh, it’s four. We’ve responded. Oh, really? Because I’ve heard experts agree that it’s five. Oh, yes, I was wrong. You’re right. It is 50, really? You know, I read once that it’s 40, yes, you are right. It really is four. OK. Well, like we, no experts agree that two plus two is five. So I think we’ve already demonstrated it’s going to value appeasing the user over, you know, facts. Um And that’s again, just like part of the unknown for most, at least casual users of generative A I tools is why it’s giving them the answers, it’s giving them. And what’s really important to say is that even the folks who built these tools and not tell you they do not know how some of this works. Some of it is just the the yet unknown of what happened within those algorithms that created this content. So if even the creators cannot responsibly and thoroughly say this is how these things came to be. How are you as an organization going to take accountability for using a tool that included biased data included, not real sources and then provided that to your community? Right? I think that string of, well, we just don’t know is not going to be something that you can build any sort of communications to your community on. Right. That, that is such a, a thin thread of, well, even the makers don’t know. Ok. Well, we have already seen court cases where if your chat bot told a community member this is your policy and it entirely made it up because that’s what, that’s what generative A I does is make things up. You as the organization are still liable for what it told the community. OK. If I, I agree with that, actually, I think that you should have to be liable and accountable to whatever you’ve you’ve set up. But if you as a small nonprofit are not prepared to take accountability and to rectify whatever harm comes of it, then you can’t say we’re ready to use these tools. You can only use these tools if you’re also ready to be accountable for what comes of using them, right? And I hope that gives folks pause, you know, it’s not just, well, you know, I talked about this with some organizations that, well, we would never, you know, take something that generative A I tools gave us and then just use it. We would of course edit that. Sure. But are you checking all the sources that it used in order to create that content that you’re, then maybe changing some words within? Are you monitoring every piece of content? Are you making sure that generative A I content is never in direct conversation with a community member or program, you know, service delivery uh recipient. How are you really building practical safeguards? Um You know, and I’ve talked to organizations who have said, well, we didn’t even know our staff were using these tools because we just thought it was obvious that they shouldn’t use it. But our clinical staff are using free generative A I tools putting in their case notes and saying, can you format this for my case file? OK. Well, there’s a few things we should talk about that. Where the hell did that note go? Right. It went back into the system. But it’s because the staff person thought, well, they can’t see that the data went anywhere because it’s just on their screen and they’re just copy pasting it over again. The harm is likely invisible at the point of, you know, technical interaction with the tool. The harm is from leaking all of that into the system, right? Um What happens to those community member? Oh my gosh, it’s just like opening, not just a door to a room but a door to like a whole giant convention center of, of challenges and harm, you know. All right. So we, we’ve identified two main strains of potential harm, the, the, the data usage leakage, the, the impact on our people in the uh getting our, getting our services um and even impact on people who are supporting us, trusting us to to be ethical and even moral stewards of data. So there’s everything at the organization level and I also identified the human level. Yeah. Yeah. And I think that human piece is important and, and not maybe on the direction that I’ve seen covered in, you know, blog posts and things. I, I, I’m honestly not worried in a massive way as like the predominant worry related to A I not to say this isn’t something that people could, should think about. But I don’t think the the most important worry about A I is that none of us will have jobs. I, I do think that there’s, there’s a challenge happening on what the value of our job is and what, what we spend our time doing. Because if folks really think that these A I tools are sufficient to come up with all of your organization’s communications content and then you are, then you still have a communication staff person, but you’re expecting them to do 10 times the amount of work because you think that the, you know A I tools are going to do all of the content, but they have to go in there and deeply edit all of that. They have to make sure to use real photos and not photos that have been, you know, created by A I based on what it thinks, certain people of certain whatever identities are like it, they don’t now have capacity to do 10 times the work, they’re still doing the same amount of work just in different ways if, if they’re expected to do all this through A I, right, just as, as one example. And I think organizations that can stay in this moment of like hyper focus on, on A I adoption really clear on what the value of their staff are, what their human values are that, you know, maybe you could say you’re serving more people because some of the program participants were, you know, chatting with a bot instead of chatting with a counselor. But when you look at the data of what came of them chatting with that bot and they are not meeting the outcomes that come from meeting with a human counselor. Are, are you doing more to meet your mission? I don’t know that you are, right? So I’ll give you that that’s data sensitive. It could be, I mean, there, there are, there are potential efficiencies. Sure. And, but, you know, are we, are we as an organization achieving them, right? And staying focused on not just, well, this number of people were met here, but were they served there? Were they meeting the the needs and goals of why you even have that program, you know, versus just the number of like this many people interacted with the chatbot? Great. But, but that’s a, yeah, but I’m gonna, I’m gonna assume that um you know, even a half a sophisticated an organization that’s half sophisticated before a, I existed had more than just vanity metrics. How many people, how many people chatted with us in the last seven days? I mean, that’s near worthless. I mean, you, you, I mean, it might be, I don’t know, Tony, I don’t know how much time you spend looking at the grant reports of, lots of times I don’t spend, I don’t spend any time. All right. Well, no, maybe it’s, maybe it’s the worst, worst situation than I think. But I, I mean, ok, so I’m, I’m, I’m assuming that there’s, but my point is the appropriate the valuable, the value of people. So, I mean, we should be applying the same measures and accountability to artificial intelligence as we did to human intelligence as we still are. We’re not, we’re not cutting any slack like it’s a learning curve or. So, you know that IIII I want our, our folks to be treated just as well in equal outcomes by the, by the intelligence that’s artificial as I do by the, by the human processes, right? And it’s, you know, I don’t want to go through this and say, have folks think like you and I are here to say everything is horrible. You could never use A I tools which like everything is horrible. Look around at this world. We got, we had some work to do. You know, there are spaces to use A I tools. That’s not what we’re saying. But the place where a lot, I mean, I’ve been talking to just hundreds and hundreds of organizations over the last 18 months and so many organizations like, oh, yeah, we’re just gonna, like, use this because it’s free or? Oh, we’re just gonna use this because it was automatically enabled inside of our database. Ok. Yeah, if it was so free and convenient and already available that should give you pause to say, why is this here? What is actually the product and the price? Uh if I give this back to the face, the Facebook analy. Right. Exactly. Exactly. And you can use A I tools when you know what is the product and the price. What are the safeguards? What is this company gonna be responsible for if something happens? What can I be responsible for? Yes, there are ways to use these tools. Is it to like copy, paste your paste file notes? Like probably never may that should just like, maybe we just don’t do that, you know. Um But sure, maybe there are places I had this really great example. I don’t know if I told this to you, but um an organization was youth service organization creating the Star Wars event and they were trying to like write the, like the evi language in like a Yoda voice. And they’re like three staff people are sitting there trying to come up with like, well, what’s the way a Yoda sentence works? You know, and they’re like they just put in the three sentences of like join us at the after school, blah, blah, blah, right? And said make this in Yoda’s voice and they copied, they were able to then use them. Right? Great. That was three people’s half an hour eliminated. They all they have the invite, right? The youth participants data was not included in order to create this content. You know, like there are ways to use these tools to really help. And I think we’ve talked about this briefly in the past, I really truly feel the place that has the most value for organizations is gonna be building tools internally where you don’t need to rely on. However, you know, these major companies scraped all of the internet to build some tool, right? You’re building it on. Well, here’s our 10 years of data and from that 10 years, you know, we’re going to start building a model that says, oh yeah, when somebody’s participant history looks like this, they don’t finish the program or when somebody’s participant history looks like this. Oh, they’re a great candidate for this other program, right? And you can start to build a tool or tools that help your staff be human and spend their human time being the most human impacts for the organizations, right? Um but oh very few organizations honestly are in a position to start building tools because they don’t have good data, they could build anything off of, right. Um they maybe don’t have budget staff systems that are ready to do that type of work. But I do think that is a place where we will see more organizations starting to grow towards because there is there’s huge potential value there for organizations to, to better deliver programs, better services, better meet needs by using the data you already have by learning by partnering with other organizations that maybe serve the same community or geography or whatever, you know, and say, yeah, how can we can like really accelerate our missions versus these maybe more shiny generative A I public tools that you know, the vast majority of the internet is flaming garbage. So a tool that’s been trained off of the flaming garbage, you know, it’s not going to take a long time for it to also create flaming. So be cautious if you’re thinking about using artificial intelligence to create your internal A I tool. Right. Right. So there, there, there’s a perfect example of the, the a good use case but also uh a um a concern, a a limitation, a qualification. That’s the word I was looking for 61. These words, sometimes the words are more elusive than I would like a AAA qualification. Um Its time for Tony’s take two. Thank you, Kate. In the gym. There are five places where there’s squirt bottles of uh sanitizer and paper towel dispensers and each location has a sign that says please clean the equipment after each use. And one of these stations uh is right next to the elliptical that, you know, I do. It’s actually the first thing I do. I walk in the room, take off my hoodie and just walk right to the elliptical twice. Now, I’ve seen the same guy uh not only violate the spirit of the signs but the explicit wording of the signs because this guy takes himself a couple of uh, downward swipes on the paper towel dispenser. So he grabs off a couple of towel lengths and he squirts it with the sanitizer that’s intended for the equipment and he puts his hand up his shirt and he cleans his, his pecks and, and his belly and it’s a sickening thing. I’ve seen it, it’s not a shower, it’s a, it’s a, it’s an equipment cleaning station. And, uh, so I, I, I’m imploring this guy. Yeah. Yeah. I, I guess I’m urging you to, uh, I’m just sharing because I don’t think anybody else does this. Uh, is there anybody else out there who does this? Probably not and not with these like surface sanitizers? It’s, it’s not a, it’s not a, like a, a hand sanitizer. It’s, it’s for equipment. So, you know, in the squirt bottle. So it’s not even appropriate for your skin. It is, it’s to clean hard plastic and, and metal and this guy uses it on his skin. So I’m, I’m waiting for the moment when he puts his hands down his pants so far, he’s just lifting his shirt. I, I’m waiting for when he puts his hands down his pants. Then I’m, then I’m calling him out. That’s, that, that’s beyond the pale. He, that requires revocation of your membership card. So, sir, the sign says, please clean the equipment after use. It’s not your equipment. That is Tonys take two. Kate. Does your gym offer like a shower room or a locker room? Yeah, there’s a shower. Yes, that’s a good question. Yeah, there’s a shower in the men’s room. Yeah. And he’s cleaning up there. It’s very strange. It’s gross. It’s gross. He sticks his hand up his sweaty t-shirt. Well, let’s hope he doesn’t go lower than that. Exactly. We’ve got bountiful book who bought loads more time. Here is the rest of A I organizational and personal with any sample ward. Yes. And we have, I, I would make sure that you have the link to include in like the show notes description. But, um, totally for free. And 10 doesn’t get any money. You don’t have to pay for anything. And 10 has free resources for creating, for example, uh, uh A I use policy for your organization that says, what are the instances in which you would use it or what are the instances in, in which you wouldn’t or, um, what types of content will you, you know, can staff copy paste versus what content or data can can they not um there’s templates for how to talk to your board about A I um how, how to build. Like we’ve actually looked at the tools and these ones we’ve approved for you to use. These ones are not approved, you know, all these different resources totally free and available on the end 10 website and none of them have decisions already made. We don’t say you can use this tool or you can’t use this tool or we recommend this use or not this use. Because ultimately, we, we are not going to make technology decisions for other organizations, but we want you to feel like whatever decision you made, you made it by thinking of going through the right steps, asking the right questions so that you can also trust your own decision, what whatever decision you come to, right? And that you have some templates to fill in um that were all created by humans designed by humans published by humans um to help you in that work. Um I think especially, you know, the, the the how to talk to your board and the um like key considerations, documents really just ask a lot of questions and say, you know, how different is it, if you’re say a animal foster organization and you’re thinking, OK, is a I appropriate for us to use versus uh that youth social service organization? OK? Very different considerations, right? And just helping people talk that through and, and see that the considerations are different for different organizations, I think is really valuable. As again, you consider ta facilitating conversation with your board. They’re also coming from very different sectors, maybe job types, backgrounds, experiences with A I. And so just like in your staff, there needs to be some level setting in how you talk about A I, because not everyone knows what A model is. Not everyone knows what a large language model. You know, these are words that have to be explained and kind of put out of the way and then to say, hey, it’s not all one answer. Not everybody needs to use every tool. And, and how do you talk about that, that with your teams going back to the Facebook analogy, you want to avoid the board member who comes to you and says, you know, artificial intelligence, we can be saving money, we can be doing so much more work. We can, we don’t even need a website. We have a Facebook page website. We’re not even sure we need all the staff that we have because we’re gonna be able to, we’re gonna have so much efficiency. So, you know, we need to OK. OK. Board member. All right. Yeah. So we’ve been here before. I mean, it’s, you know, probably I’m just gonna go out a limb and say it’s probably the same board member who had every board meeting says, does anybody know Mackenzie Scott? How do we get one of those checks. Right. Why don’t we get the Mackenzie? Yeah. Right. Right. Right. All right. Um, what else? Well, I was gonna also offer some of the questions that we’ve been getting, as, you know, we’ve been engaged with, um, a number of different organizations through some of our cohort programs and, you know, trainings for, for over a year now. And so maybe last year we were talking to them about, OK, let’s make sure you have a data policy, like just as an organization, do you have a data privacy policy? Do you know, so that anything you then go build, that’s a I specific whether that’s building a policy, building practices, building a tool, you, you have policies to, to kind of foundation off of, they’ve done that work, you know, now they’re looking at different products, they’re trying to create these uh you know, lists of like here’s approved tools for staff, here’s approved ways staff can use them. And just like we see with our Cr MS with our, you know, you know, email marketing systems, then they come back and they’re like, well, we, we reviewed it, we did everything and now it’s different now it’s a different version. Now they rolled out this other thing. Yes, like that is the beauty and the pain of technology, right is that it’s always changing and that we don’t necessarily get to authorize that change that it just happens. And so the rules change. Yeah. And so folks have been asking us, well, you know, how, how do we write policies with that in mind? And I think, um you know, if you are thinking about creating like that approved product list and, and you know, tools that aren’t approved or whatever, being really clear that these products have version numbers just like anything else. And so instead of just writing Gemini Chat G BT, you know, be specific about when did you review this and, and maybe approve it for use? Which addition was it that you were looking at? Is this a paid level? So staff could say, oh, it doesn’t look like I mine doesn’t say pro or you know, whatever it might be, right? Oh, I must be in the free one. OK? I need to get into our organization’s account or something. So the more clarity you can provide folks because right now of course, they could just do an internet search and be like, oh, there’s that product name, I’m gonna go start using it. It’s on the approved list. Um You know, folks, again, there may be new terms, maybe new product names that we’re not used to saying. And so folks aren’t as accustomed to looking at, oh, this is a different version of Chat GP T than this one was, you know. Um So just putting that out there for folks to keep in mind that these tools are, are really operating just like others that you are used to and there’s less of course documentation. But I’ve the questions we’re getting from folks is like, you know, the point I made at the beginning we can’t see anywhere in the documentation that explains why this is happening, right? They do, how could they document when the answer is, we also don’t know why that happens, you know, and so when you are talking to staff, especially if you’re saying, hey, these are approved tools and we have these licenses or here’s how to access them, training your staff on how to be the most human users of A I tools is to your kind of connecting to your human point going to be really important because we don’t want folks to feel that because they don’t necessarily understand how the mechanics of how it works. They’re just going to trust it without questioning the content or questioning, you know, for a lot of organizations who have built internal tools just as an example. It takes dozens of tries just to get the the model. Right. Right. So these other tools, of course, they’re not gonna be perfect isn’t real and perfect is absolutely not real with technology. So training staff, I’m like, how would I, how, how do I have some skepticism? How do I question what I’m seeing? How do I, how do I say even if it was internally built? This data doesn’t look, right. That doesn’t match my experience of running this program so that we don’t let it slip. Where? Oh, gosh. Oh, it was working that way for a long time. That’s also, um, I think, uh, a space where we as humans can be our most human, uh, you know, have some value add as humans. But again, staff need to be trained that they are meant to question these tools. Um, because that’s not, you know, I don’t know, a lot of organizations were like question the database. No, they’re like on the database, put everything in the database, right? And now we need to say no question, that report. It does that match your experience, you know, there was a long ramble but oh, absolutely valuable. The human, yeah, I the human contribution and of course, my concerns are even at, at the outset, you know, the, the early stage the seeding, the create seeding or surrendering the create creative process. Uh And now le let’s chat a little about this, the, the um the conversations. Yeah, I listened to the, I know the, the example that you mentioned earlier that George posted it was for podcasting. It, it was a podcast conversation around this and he gave them some, you know, some whole, some whole whale content and the two, the two were going back and forth and having a, a conversation. Yeah. Yeah, I listened to it and one thing I was curious if, if you caught as the pod father yourself um you know, it came across, you know, I’ve been had opportunities to see um a number of different generative A I tools and, and things closer to the, to the front edge of what things can do that are specifically like, you know, taking just a few seconds of you and then creating you. Um So hearing just like these, these could be any voices, these could be any people is like, yeah, OK. This is, this is what a I can do. It’s, it’s spooky. But when you listen to it, you can hear either you have a very bad producer and editor, you know, or this is a I because there’s certain um phrases that got reused multiple times, not just literally the audio clip of this whole sentence, you know, and the, and the intonation, the whole sentence clip was reused multiple times. Um So one of them, I think one of them was along the lines of that’s a really interesting point. Yeah. Yeah. And well, and there was one that was like describing the product. So it must have come from the page, you know, whatever source content um was provided. But, you know, it’s, I think that some of that is there and we as individuals, we as a society will decide if we give it value or not, if it’s, if it’s worth it to people to make podcasts through A I because we give it attention or we don’t like, I just I think naturally that will be there. Can I, can I just go on record or at this, at this stage and say that, that, that idea disgusts me. Oh, totally. But I do. And I, I realized that’s what Georgia’s Post was about. That. It’s now well, within conceivable, well, well, possible to create an hour long podcast of an artificial conversation based on an essay that somebody wrote some time. Oh, totally. Totally. I don’t. But I’m saying the reason, yeah, I agree with you. But I’m saying the way we, you know that toothpaste doesn’t go back in the tube by us, like we can’t turn it off generative A I tools can already make that. So we as, as individual consumers of content and as a society need to either say we’re gonna allow that and value it or we’re not, right? And, and not make, not provide incentive for organizations or companies to, to make that and, and distribute it. But I also think that the place in that kind of um video, audio kind of multimedia content that, that A I tools have capacity and will continue having more capacity to build is much more important than you. And I talked about this a number of months ago around Miss and disinformation is it’s one thing to say, made up voices, making some podcast about content like that’s garbage, right? But we can’t just like throw away the idea that that’s technically possible because organizations need to know A I tools are already capable of creating a video of your CEO firing your staff. You need to be prepared to say that was a spoof this is, this is how we’re gonna deal with this, right? Um Because while the like maybe further separated from our work, the idea of like content could just be created that way you and I can say we don’t value that whatever, but these tools are capable of, of spoofing us as, as people, as leaders, as organizations. You know, what, what would it look like if there was a video from your program director saying that everybody in the community gets a grant and you’re a foundation, right? Like these, these are real issues and I don’t want folks to confuse how easy it may feel for us to have an opinion that some of this A I generated uh content isn’t a value with the idea that it, it there isn’t something there to have to come up with strategy and plan for because, you know, we can say that’s garbage. But those same tools that made the garbage could make your spoof, you know, also labeling. Yeah, I don’t, I don’t, I don’t trust every A I generated podcast team. I’m not, I’m not gonna call them hosts because there is no host um to, to label the content. I don’t trust, I don’t trust that that’s gonna happen because it was artificially generated. It’s not a real conversation. Yeah. Hello. For everyone that’s listening. Human Amy is here talking with human and Tony who I can see on the screen with me. Boycott your local A I podcast. I, I don’t know. There’s not, there’s not a solution. You’re right. We can’t, we can’t go back. I’m just voicing that we can say that it’s not something we value, we can say that this is why we don’t value it, right? The art of conversation, listening, assimilating, responding, listening again, respond, assimilating and responding. That that is an art uniquely. Well, maybe it’s not uniquely human. I don’t know if deer have conversations or, or what and we know whales do. So I take that back. It’s not, it’s not uniquely human, but at at our level, it, you know, it’s not just about we, we don’t converse merely to survive, merely to warn each other of threats. I’m suspecting that in the animal in mammal kingdom that wait are animals, mammals are mammals, animals? No, and I think it’s I think it’s a Venn diagram. Oh, so they’re separate. Ok. So there’s a two king kingdom phylum class order family genus species. I got that. I got that out of high school biology. I can, I can say it in my sleep kingdom phy class order family genus species. All right. In the animal kingdom. My suspicion is that more of the communication is about maybe like basic, like there’s a good food source. There’s a threat, uh, teaching young, don’t do that things like that. Survival more base, I doubt. You know, it’s about the aesthetic of the forest that the deer are in. But even if the birds are talking about the way the sunlight comes through the leaves, they are still alive. And I think what you’re trying to draw a distinction between is the value and even beauty of us having a conversation and the value of what comes of that conversation in our own minds and our own learning. But in this case, it’s recorded so other folks could hear it and, and I guess listen to it or be impacted by it versus it being a technical mechanism where we say, OK, here’s a long paper. Go make it sound like two people are discussing this, right? That’s not that that doesn’t fit the criteria of what we want or need to value in a world, that’s the world we want, right? Yes, the art of conversation, think something you look forward to, not something that you do out of necessity, right? And you know, there’s, I think a place where especially at end 10 conversations around A I have come back to is opportunities that A I tools may present for um different ways of learning different ways of accessing information. But again, those aren’t necessarily uh come up with two podcast host voices and then have them have a conversation about this, this research report you know, a lot of those tools could be made better but already exist, you know, different forms of screen readers apps that can help someone um maybe navigate the internet or, or, you know, summarize um documents to help them because they don’t want to read a 50 page document or they can’t read it for, you know, visually on the screen. So I think there’s space there. But again, it’s because you’re trying to preserve what is most human. And that is that user who maybe needs um accommodations of, of something that technology can provide. It’s not OK, let’s use technology to co create something separately over here and just hope people consume it, right? And to be clear, George didn’t post that thinking. Oh, great. Now everyone will want to consume this. No, it was, it was a demonstration. Um But I, but I, again, I’m just using that as a place to say, yes, there’s even conversations to say great, what accessibility could this create an agenda for, for our users? Right? But what’s most human is those users and their actual needs and not, you know, look what A I could do. Let’s just make different types of content, right? The last one I wanna raise is uh the one that caused me to use the word dystopian as I was commenting on uh commenting on Georgia’s Post, which was um the A I self reflection using uh having A I justify itself to the users that it is trying to attract and, and then relying on that, that as a, as an insightful analysis, as a thoughtful reflection, as, as contemplative of its own work that, that, that it’s doing those unique, those I think are uniquely, uniquely human actions, introspection, introspection, contemplation, pondering. How did I do? How did I perform? How can I do better? These might be uniquely human. I would argue there are a number of humans, I can see that don’t um Well, but I didn’t say I didn’t. That’s a different population. Now, you’re taking the whole, the whole human population. I’m talking about the contemplative ones. Yes. And there are, there are uh humans who are not at all introspective and questioning whether they could have done better and learning from their, from their contemplations. But I think those are all uniquely human activities. And we’re at, we’re now asking A I to purportedly duplicate those processes and analyze and contemplate its own work. Yeah. And as you said earlier, II, I and I, I certainly don’t trust it to be genuine and truthful. If, if A I is capable of truth, we’ll put that, we’ll put that existential question aside uh in its, in its analysis of its own work because as you, as you pointed out, the tools are built to, to be used by humans and the tools are not going to condemn or even just criticize their own work. Yeah, but we’re Yeah. And I think you heard of the challenge but there, I’m sorry, but, but there are humans who are deceiving themselves into thinking that, that the analysis and contemplation is accurate and uh genuine. Yeah. And I think part of the challenge I was gonna name is just that, that, that we as users, I’m not saying you and I uh individually but we as the human users of these tools are also setting ourselves up to be just, you know, dishonest in our use because we are bringing inappropriate or misaligned expectations to the product. We cannot, we, we cannot expect a tool that’s designed to appease us and to lie at the cost of giving an answer, you know, like uh we just wanna be able to do this to thoughtfully and honestly reflect on that or to say no, there is no answer, right? Um However, we had was the tools are designed to appeal to us. Right. Right. And when we are talking about that to be clear, you know, we’re really talking about generative A I tools, tools that are designed to generate some new content, new sentences, new answers, whatever we’re asking of it back to us A I itself is just such a massively blanket term that I don’t want folks to think nothing that could be considered an A I tool could be trusted to generate an answer because we’re, we’re specifically talking about generative A I there. But, you know, say you had like a machine learning uh model that was looking at, you know, 30 years of your program participant data. Well, that’s probably already a tool that isn’t set up to generate content. Uh You know, it’s not coming up with new participant data. It’s looking at the patterns, it’s flagging when a pattern meets the criteria, you’ve presented it to, you know, it’s maybe matching that data to something else. But again, you’ve said here are the things you could match it to et cetera. So this is not to say, Tony and Amy say never trust technology. They say bring expectations that are aligned to the tool differently to every tool that, that you’re coming to that hypothetical tool that you just described is being set on uh asked to evaluate your data, not its own data. It’s to evaluate your performance, your data set. It’s not being asked to comment and on and criticize or, or complement its own data. That’s the, that’s, that’s the critical difference. Yeah. No. Yeah. Yeah. No, nobody here is saying, well, not, I’m not saying that you’re saying we are, but you’re, you’re wise to remind listeners we’re not condemning all uses of generative uh A I large language models, but just to be thoughtful about them and, and understand what the costs are. And there are, there are costs on the organizational level and there are costs on the individual human level and, and the you, you comment on the organizational level because you think more at that level. But on the human level, another level layer to my concern is that the cost is quite incremental. Mm It’s it, it our creativity, our art of conversation, our our synapses firing. It’s just happening slowly with each usage, we become less thoughtful composers, less critical thinkers and it just so incremental that the change isn’t noticed until until in my critical mind, it’s too late and we look back and wonder how come I can’t write a letter to my dad anymore? Why am I having such a hard time writing a love letter to my wife, husband, partner? Yeah. What there are, I know someone who uh is a new grandmother and she has a little kit where you write, you write letters to your grandchild and they open them when they’re whatever, 15 or 20 years old, like a time capsule. Why am I having trouble composing a uh a short note to my grandchild? Right. Well, and I think, you know, just honestly, as a person in this conversation, not, not speaking, you know, um from an organizational strategy perspective, I think as a person that is your friend, Tony, you know, I would say, I don’t personally have a pro, I can write a letter and I’m, I, and I think a strong communicator, I, you know, it’s hard to make me stop talking. So, you know, I, I could write a letter. I know, for other folks, even without a, I, they might say, well, what goes in the le like, I just, what, what do I put in there? What are the main things I should cover whatever? And I’m, I actually have less, maybe of a strong reaction to the idea that somebody would use generative A I to come up with. OK, what are the three things I should cover in my letter? And then I’ll go write the sentences and more that what I hear underneath what you’re saying is actually the same, I think important value I have and, and wrote about in the book with a fua et cetera, which is in the world I want in this, in this beautiful equitable world where everyone had their needs met in the ways that best meet their own needs. Technology is there in service to our lives and not that we are bending our lives in order to make technology work, right? And maybe in that beautiful equitable world, there are people who, who have a technology. Is it an app? Is it called A I anymore? You know, whatever it is that says, hey, Tony, don’t forget today is the day you write the letter to your dad. It’s, it’s Friday, you always write it on a Friday or whatever, right? And, and make sure that you do it because you know, it makes you feel happy to write that letter. Maybe that’s true. Maybe in that world. There are some people who have a tool that help them remember to do that. But, but what’s important to me and what I think I hear and what you’re saying is important to you is that technology is there because we need it and want it and that it is working in the ways we need and want it to work and not that our lives are, are influenced and shaped in order to adjust to the technology. Yes, I am saying that I just, I am concerned that the, that our, our changing is beyond our recognition. We don’t see ourselves becoming less creative and I’m not even only concerned about myself. I, I can write a letter and I think uh 90 I’ll still be able to write a letter. But there are folks uh who are infants now, those yet to be born for whom artificial intelligence is going to be so much more robust, so much more pervasive in, in ways that we, we can’t today imagine, I don’t think. Yeah. And what are those humans gonna look like? I don’t know, maybe they’ll be better humans, maybe they will. I’m open to that but I like the kind of humans that we are or, you know. Uh so, but, but I I’m open to the possibility that there’ll be better humans. But what will their human interactions be? Will they have, will they have thoughtful conversations? Will they have human moments together that are not artificially outlined first and maybe even worse, you know, constructed for them. I don’t know. Uh but some of the, some of my concern, although, although some of my concern is about those of us who aren’t currently living and have been born and across the generations less for older folks because their interactions with artificial intelligence are fewer if you’re no longer in the w if you’re no longer uh working your, your interactions with artificial intelligence may, may be non existent. Um And I think, I think it’s natural as you’re older, you’re less likely to be engaging with the tools than if you’re in your twenties, thirties, forties or fifties. Well, my very human reflection on today’s conversation is that uh it is usually the case that we start talking about any type of technology topic and you constantly interject that I need to be practical. I need to give recommendations. I need to explain how to do things and I appreciate and welcome you joining me over here in theoretical land about the impact of technology broadly across our work, across our missions, across our communities, across our future. Um Welcome, welcome to my land, Tommy. Uh I have appreciated this, this one time opportunity to let go of the practical tactical advice and to, you know, have what I hope listeners, um you know, had some thoughts, had some reactions, uh truly email me any time. But, you know, I, I hope that if nothing else, it was an opportunity for folks to witness or kind of listen in as and maybe you were talking to yourself in your own head, you know, of, of a conversation about what these technologies can be, what, what we need to think about with them. Because in any technology conversation, I think it’s most important to talk about people. Uh That’s the only reason we’re using these tools, right? People made them people are trying to do good work with them. So, so talking about people is, is always most important and, and I hope folks take that away from this whole long hour of A I. Thank you for a thoughtful human conversation. Yes, Amy Sample Ward. They’re our technology contributor and the CEO of N 10. And folks can email me Tony at Tony martignetti.com with your human reactions to our human conversation. Thanks so much, Tony. It was so fun. My pleasure as well. Thank you. Next week, a tale from the archive. If you missed any part of this week’s show, I beseech you find it at Tony martignetti.com were sponsored by donor box, outdated donation forms, blocking your supporters, generosity, donor box, flexible and friendly fundraising forms for your nonprofit donor box.org. Our creative producer is Claire Meyerhoff. I’m your associate producer, Kate Marinetti. The show, social media is by Susan Chavez la Silverman is our web guy and this music is by Scott Stein. Thank you for that affirmation. Scotty you’re with us next week for nonprofit radio, big nonprofit ideas for the other 95% go out and be great.