logo
Keep the Peace panel focuses on community issues

Keep the Peace panel focuses on community issues

Yahoo18-05-2025
CONNEAUT — The Conneaut Area Ministerial Association and the Jefferson Community Church of God in Christ hosted a Keep the Peace panel at First Congregational United Church of Christ Friday.
The panel included Ashtabula County Sheriff William Niemi, Conneaut Municipal Court Judge Nicholas Iarocci and county Mental Health and Recovery Services Board Executive Director Kaitie Park Hart.
The panel was facilitated by Reverend Isaac Chappell, and discussed community outreach and dealing with people suffering from addiction or mental health issues, and how their institutions relate to it.
'Prevention starts in the family and in the community,' Park Hart said. 'There is only so much we can do with prevention in our schools that's actually going to impact our kids.'
Park Hart said the MHRS has a mobile crisis team that works with law enforcement.
'Whenever there is a crisis situation, law enforcement can call our mobile crisis provider, which is Signature Health,' she said. 'Unfortunately, it's not 24/7. We're working on that. Workforce is a huge issue in Ashtabula County. For all of the services that we have available, we need more people to provide those services.'
Niemi talked about county leaders' interest in getting funding to expand the jail.
'That jail was built in 1977,' Niemi said. 'The people that we deal with now are on a whole different level. Mental health is one of the biggest issues we're dealing with today.'
Park Hart said the MHRS board works to provide services at the county jail.
'There are a couple therapists in the jail,' she said. 'We have a case manager. Like the sheriff said, space is absolutely an issue.'
Niemi said the jail lacks space for counselors.
'Our jail is falling apart,' he said. 'It doesn't meet our needs, basically, and it's very possibly costing the county over three and a half million dollars.'
Iarocci talked about the Conneaut Renewal Court he started in the city. It is one of three Substance Abuse and Mental Illness courts in Ohio, and focuses on helping people suffering from those issues, who are sent to the municipal court.
'There were two people in [the Conneaut] jail, our little jail, who had serious mental health issues and substance abuse issues,' he said. 'One young lady was sleeping on the streets and was 23 years old.'
Iarocci said he wanted the 23-year-old woman sent to Renewal Court, regardless of what happened with her case.
'When she was sentenced ... I basically said we have a bed for you for the next hour, and you're immediately going to residential treatment,' he said.
The woman had never had residential treatment before, he said.
'She just started crying,' he said. 'This was a girl, who three days before, refused to be assessed.'
Reverend Eddie Parker, a community engagement liaison, was also part of the panel.
Parker said churches can get involved with community outreach and helping youth suffering from substance abuse issues.
'We're not proselytizing and trying to engage them in faith,' he said. 'We're trying to be a friend and meet the emotional needs. The church happens just to be a safe, neutral meeting place.'
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

The Real Risks of Turning to AI for Therapy
The Real Risks of Turning to AI for Therapy

WebMD

time3 hours ago

  • WebMD

The Real Risks of Turning to AI for Therapy

Aug. 20, 2025 — Whenever Luke W Russell needs to work through something, they turn to ChatGPT. (Luke uses they/them pronouns.) 'I've wept as I've navigated things,' said the Indianapolis filmmaker, who uses the chatbot to pick apart intrusive thoughts or navigate traumatic memories. 'I've had numerous times when what ChatGPT is saying to me is so real, so powerful, and I feel so deeply seen.' Russell's experience reflects a broader, growing reality: Many people are turning to chatbots for mental health support — for everything from managing anxiety and processing grief to coping with work conflicts and defusing marital spats. More than half of adults ages 18-54 — and a quarter of adults 55 and up — say they would be comfortable talking with an AI chatbot about their mental health, according to a 2025 survey by the Harris Poll and the American Psychological Association (APA). The catch: OpenAI's ChatGPT and other chatbots — like Anthropic's Claude and Google's Gemini — are not designed for this. Even AI products promoted as emotional health tools — like Replika, Wysa, Youper, and MindDoc — were not built on validated psychological methods, said psychologist C. Vaile Wright, PhD, senior director of the APA's Office of Health Care Innovation. 'I would argue that there isn't really any commercially approved, AI-assisted therapy at the moment,' said Wright. 'You've got a whole lot of chatbots where there is no research, there's no psychological science, and there are no subject matter experts.' Critics warn that AI's potential for bias, lack of true empathy, and limited human oversight could actually endanger users' mental health, especially among vulnerable groups like children, teens, people with mental health conditions, and those experiencing suicidal thoughts. The growing concern has led to the emergence of the terms 'ChatGPT psychosis' or ' AI psychosis ' — referring to the potential harmful mental health effects of interacting with AI. It's even drawing attention from lawmakers: This month, Illinois enacted restrictions on AI in mental health care, banning its use for therapy and prohibiting mental health professionals from using AI to communicate with clients or make therapeutic decisions. (Similar restrictions have already been passed in Nevada and Utah.) But none of this is stopping people from turning to chatbots for support, especially amid clinician shortages, rising therapy costs, and inadequate mental health insurance coverage. 'People have absolutely reported that experiences with chatbots can be helpful,' said Wright. The Draw of Chatbots for Mental Health Data shows we're facing a massive shortage of mental health workers, especially in remote and rural areas, said psychologist Elizabeth Stade, PhD, a researcher in the Computational Psychology and Well-Being Lab at Stanford University in Stanford, CA. 'Of adults in the United States with significant mental health needs, only about half are able to access any form of treatment. With youth, that number is closer to 75%,' said Jessica Schleider, PhD, a child and adolescent psychologist at Northwestern University in Chicago. 'The provider shortage is clearly contributing to why so many folks are turning to their devices and, now increasingly, to generative AI to fill that gap.' Unlike a therapist, a chatbot is available 24/7. 'When [people] need help the most, it is typically after hours,' said Wright, who suggested the right AI tool could potentially supplement human therapy. 'When it's 2 a.m. and you're in crisis, could this help provide some support?' Probably, she said. Results of the first clinical trial of an AI-generative therapy chatbot showed 'significant, clinically meaningful reductions in depression, anxiety, and eating disorder symptoms' within four to eight weeks, said lead study author Michael V. Heinz, MD, a professor at Dartmouth College's Geisel School of Medicine and faculty affiliate at the Center for Technology and Behavioral Health in Lebanon, New Hampshire. The chatbot — Therabot, developed at Dartmouth — combines extensive training in evidence-based psychotherapy interventions with advanced generative AI. 'We saw high levels of user engagement — six-plus hours on average across the study,' Heinz said. Participants said using Therabot was like talking to a human therapist. But results are early, and more studies are needed, Heinz said. Access and affordability drew Russell to ChatGPT, they said. 'I didn't set out to use ChatGPT as a therapist. I quit therapy in January due to income dropping. I was already using ChatGPT on the regular for work, and then I started using it for personal idea exploration. ... I've never had a therapist who could move as fast as ChatGPT and ignore miscellaneous things,' they said. Perhaps one of the most appealing aspects is that chatbots don't judge. 'People are reluctant to be judged, and so they are often reluctant to disclose symptoms,' said Jonathan Gratch, PhD, professor of computer science and psychology at the University of Southern California, who has researched the topic. One of his studies found that military veterans were more likely to share PTSD symptoms with a virtual chatbot than in a survey. When Chatbots Are Harmful Most people don't know how AI works — they might believe it's always objective and factual, said Henry A. Willis, PhD, a psychologist and professor at the University of Maryland in College Park. But often, the data they're trained on is not representative of minority groups, leading to bias and technology-mediated racism, Willis said. 'We know that Black and brown communities are not adequately reflected in the majority of large-scale mental health research studies,' Willis said. So a chatbot's clinical symptom information or treatment recommendations may not be relevant or helpful to those from minority backgrounds. There's also an impersonal aspect. Chatbots do what's called ecological fallacy, said H. Andrew Schwartz, PhD, associate professor of computer science at Stony Brook University in Stony Brook, NY. They treat scattered comments like random data points, making assumptions based on group-level data that may not reflect the reality of individuals. And who's accountable if something goes wrong? Chatbots have been linked to cases involving suggestions of violence and self-harm, including the death of a teen by suicide. Some chatbots marketed for companionship and emotional support were designed with another incentive: to make money. Wright is concerned that they may unconditionally validate patients, telling them what they want to hear so they stay on the platform — 'even if what they're telling you is actually harmful or they're validating harmful responses from the user.' None of these conversations are bound by HIPAA regulations, either, Wright pointed out. 'So even though they may be asking for personal information or sharing your personal information, they have no legal obligation to protect it.' The Psychological Implications of Forming Emotional Bonds With AI In an opinion article published in April in the journal Trends in Cognitive Sciences, psychologists expressed concern about the long-term implications of forming emotional bonds with AI. Chatbots can replace users' real relationships, crowding out romantic partners, co-workers, and friends. This may mean that individuals begin to 'trust' the opinion and feedback of chatbots over real people, said Willis. 'The ongoing positive reinforcement that can happen instantly from interacting with a chatbot may begin to overshadow any reinforcement from interacting with real people,' who may not be able to communicate as quickly, he said. 'These emotional bonds may also impair people's ability to have a healthy level of skepticism and critical evaluation skills when it comes to the responses of AI chatbots.' Gratch compared it to hunger and food. 'We're biologically wired to seek out food when we get hungry. It is the same with social relationships. If we haven't had a relationship in a while, we may feel lonely, and then that motivates us to go out and reach out to people.' But studies suggest that social interaction with a computer program, like a chatbot, can sate a person's social needs and demotivate them to go out with friends, he said. 'That may have long-term consequences for increased loneliness. For example, research has shown people who compulsively use Facebook tend to be much more lonely.' Counseling with a therapist involves 'a natural curiosity about the individual and their experiences that AI cannot replicate,' Willis said. 'AI chatbots respond to prompts, whereas therapists can observe and ask clinical questions based on one's body language, a synthesis of their history, and other things that may not be conscious to the client — or things the client may not even be aware are important to their mental health well-being.' The Future of AI Therapy "I think there is going to be a future where you have really well-developed [chatbots] for addressing mental health that are scientifically driven and where they are ensuring that there are guardrails in place when somebody is in crisis. We're just not quite there yet,' said the APA's Wright. 'We may get to a place where they're even reimbursed by insurance,' she said. 'I do think increasingly we are going to see providers begin to adopt these technology tools as a way to meet their patients' needs.' But for now, her message is clear: The chatbots are not there yet. 'Ideally, chatbot design should encourage sustained, meaningful interaction with the primary purpose of delivering evidence-based therapy,' said Dartmouth's Heinz. Until then, don't rely on them too heavily, the experts cautioned — and remember, they are not a substitute for professional help.

Behavioral Activation Therapy Cuts Perinatal Suicide Risk
Behavioral Activation Therapy Cuts Perinatal Suicide Risk

Medscape

time7 hours ago

  • Medscape

Behavioral Activation Therapy Cuts Perinatal Suicide Risk

TOPLINE: In a study of 1117 perinatal adults receiving behavioral activation therapy, the odds of endorsing suicide ideation decreased by 25% with each treatment session and by 80% at 3 months post-randomization, with specialist and nonspecialist providers showing equal effectiveness. METHODOLOGY: A multisite, noninferiority, four-arm randomized clinical trial called the Scaling Up Maternal Mental healthcare by Increasing access to Treatment (SUMMIT) compared clinicians (nonspecialist vs specialist) and modalities (telemedicine vs in-person) in delivering behavioral activation therapy. Research was conducted at university-affiliated networks in Chicago, Illinois; Chapel Hill, North Carolina; and Toronto, Ontario, Canada. Participants included pregnant (≤ 36 weeks) and postpartum (4-30 weeks) adults with depressive symptoms (Edinburgh Postnatal Depression Scale score ≥ 10). A total of 1117 participants who completed at least one treatment session and provided ≥ 1 week of Edinburgh Postnatal Depression Scale data were included in the analysis. Treatment consisted of a manualized 6- to 8-session perinatal behavioral activation intervention delivered weekly. TAKEAWAY: Among 1230 enrolled pregnant and postpartum adults, 1117 completed at least one treatment session, with 264 (23.6%) endorsing suicide ideation during treatment. The odds of endorsing suicide ideation decreased by 25% with each additional treatment session (odds ratio [OR], 0.75; 95% CI, 0.58-0.96; P = .03). At 3 months post-randomization, the odds of endorsing suicide ideation decreased by 80% compared with any time during treatment (OR, 0.20; 95% CI, 0.14-0.27; P < .001). No significant differences were found in the odds of endorsing suicide ideation between clinician types (nonspecialist vs specialist) or delivery modalities (telemedicine vs in person). IN PRACTICE: 'Behavioral activation is a first-line recommended treatment for perinatal depression and may reduce postpartum suicide risk by increasing values-consistent living and awareness of ineffective behaviors,' wrote the authors of the study. SOURCE: The study was led by Parisa Kaliush, PhD, Department of Psychiatry, School of Medicine, University of North Carolina at Chapel Hill, and Daisy Singla, PhD, Centre for Addiction and Mental Health in Toronto. It was published online in JAMA Psychiatry. LIMITATIONS: High suicide risk prompted exclusion from SUMMIT trial, and the final sample was highly educated, suggesting some limitations in generalizability. However, the prevalence of suicide ideation at baseline (15.0%) was higher than previously cited rates. DISCLOSURES: This study was funded by the Patient-Centered Outcomes Research Institute. Bradley Gaynes, PhD, disclosed receiving compensation for authorship/review of UpToDate chapters on depression. Samantha Meltzer-Brody, MD, MPH, reported receiving research funding to the University of North Carolina for clinical trials sponsored by Sage Therapeutics, Electromedical Products International, and Sirtsei Pharmaceuticals, serving as a clinical advisor and professional corporation owner for Modern Health, and serving as a scientific advisor to EmbarkNeuro and Seaport Therapeutics. Simone Vigod, MD, disclosed receiving royalties from UpToDate for authorship of materials on depression and pregnancy. This article was created using several editorial tools, including AI, as part of the process. Human editors reviewed this content before publication.

Do Phones Really Wreck Kids' Lives?
Do Phones Really Wreck Kids' Lives?

Yahoo

time8 hours ago

  • Yahoo

Do Phones Really Wreck Kids' Lives?

I am addicted! To my phone. I check an email and before I realize it, I'm watching TikTok videos: lions fight hyenas, military dads reunite with kids, athletes do amazing things. I look up, and an hour has passed. I've wasted time, ignored my family and friends, and accomplished nothing. But who cares? I'm old. I've already achieved what I'm likely to achieve. Still, what about kids? "Attention spans are declining," says psychologist Jonathan Haidt. "Levels of anxiety, depression, self-harm were pretty stable…all of a sudden, the rates go way up, especially for girls." His bestselling book (on bestseller lists for more than a year!) blames smartphones. "Once they get a smartphone…time with friends plunges. One of the best things you can do as a kid is hang out with friends, joke around, have adventures. If your kids went through puberty on a smartphone with social media, they came out different than human beings before that." My son, Max, once worked for social media companies. Now he makes his living speaking to students about how phones hook them. He compares smartphones to casino slot machines. "All the things we love about social media, those are the reward in the slot machine…we get that 'hit' once in a while….That's there to keep us scrolling for hours." Haidt agrees, calling smartphones a "gambling machine." They say some apps are worse than others. "Instagram, Facebook, Snapchat, TikTok. Those really shatter attention spans. In terms of exposure to things that are really dangerous, Snap is the worst," says Haidt. "In terms of destroying your ability to pay attention, TikTok is the worst. In terms of destroying a teenage girl's sense of confidence, self-esteem, body image, Instagram is the worst." He says social media affects boys and girls differently. "Check in on the kids at age 14, girls are doing worse. They're more depressed and anxious, more messed up." But a few years later, he says, "Girls are more likely to have gone to college, gotten a job, and moved out of their parents' home. Boys are more likely to still be in their parents' basement playing video games. They never grew up. Real life is incredibly boring compared to a video game or porn." Teachers say phone addiction makes it harder to teach. "When you and I were in school," says Haidt, "suppose they let you take your TV into class. You couldn't possibly learn." These are big problems, but I'm a skeptic. Do phones really wreck kids' lives? We don't know that, say researchers like psychologist Chris Ferguson. "Correlation does not equal causation." "But teen depression is skyrocketing," I push back, "up 145 percent for girls since 2010!" "Teen suicide was actually very high in the early 1990s," he replies, "then it decreased…way before social media….Dr. Haidt has cherry-picked a lot of data and presented only the data that support his narrative." "I am not cherry-picking!" Haidt replies. "I'm the only one in this debate who has picked all of the cherries and laid them out on a blanket." He does lay out alternative possibilities, like teen marijuana use and the decline of marriage. "My theory is the only plausible one out there," he says. "No one's even proposed one that will work across so many countries. When you ask people to get off of social media for more than a week, their levels of depression, anxiety, go down." His book suggests that parents ban phones until high school. I push back. Kids will complain, "All my friends have one!" "But what if it was only most of your friends?" he replies. "Then it's much easier." He wants schools to ban phones, and many have. I ask Ferguson, "What's the cost of banning it in schools?" "Unintended negative consequences," he replies. "Are we suspending kids for cellphone use? A lot of schools are, and that can cause real harm to the kids." Haidt insists, "When schools ban phones, the results are overwhelmingly positive….Kids know that life would be better if they didn't spend five or six hours a day on social media. They know that, but they can't help it." COPYRIGHT 2025 BY JFS PRODUCTIONS INC. The post Do Phones Really Wreck Kids' Lives? appeared first on Solve the daily Crossword

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store