logo
Penn-Trafford School District dismisses 2 students over residency concerns despite mom's objections

Penn-Trafford School District dismisses 2 students over residency concerns despite mom's objections

CBS News23-07-2025
A mother said her two kids were dismissed from school for the upcoming year because the Penn-Trafford School District said they no longer live there.
Jennifer Brown is a single mother of two boys, Landon and Lorenzo, who she says live with her at a house on Burrell Hill Road in Penn Township.
"My father owns it," Brown said, referring to the house. "My Grandfather built it. I will be inheriting it next."
But in a letter sent out by the Penn-Trafford School District on July 9 to Brown's Penn Township address, she was informed that the school district didn't believe that she lived where she said she did.
The letter went on to say that they were dismissing Landon from Penn-Trafford High School and Lorenzo from Harrison Park Elementary.
Brown did what most mothers would do and started communicating with Superintendent Dr. Matthew Harris.
The superintendent's office told her that someone reported her sitting in her car at the top of her driveway when the buses were picking up and dropping off, something that is not illegal or an indication for or against residency.
"I voluntarily gave them a tax return, an internet bill, a pay stub, my welfare information, my vehicle registration, the titles to my car, my car insurance," Brown said. "Anything that I could find with my name and address, I have given to them, and my children are still currently withdrawn from the district."
Brown was even asked to send guardianship paperwork to the school district, but she refused to do so because, one, she says she is the birth mother of her two sons, and two, she believes it would be illegal for her to present herself as a guardian and not a parent.
Brown also told KDKA that while she has lived in other places in her life, this is where she has been living for a while now. She also said that she has an older daughter named Monica who just graduated from Penn-Trafford High School, and this was never an issue while she was in school.
Brown says that this has been incredibly stressful on her and her family.
KDKA was able to see one of Brown's tax forms from last year, and it does have her name and that Penn Township address, plus it lists Landon and Lorenzo as her dependents.
KDKA called the Westmoreland County Courthouse and confirmed that her father's name is the name on the deed to the house where she says she's living.
In a statement to KDKA, Harris said,
"We take residency requirements very seriously at Penn-Trafford and we have conflicting information regarding this concern, or we would not be looking further into the situation as we are currently in the process of doing right now."
When asked what she would like to see happen, Brown said, "I hope that my children would be re-enrolled and that someone could issue me an apology."
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

AI Is a Mass-Delusion Event
AI Is a Mass-Delusion Event

Yahoo

time3 hours ago

  • Yahoo

AI Is a Mass-Delusion Event

The Atlantic Daily, a newsletter that guides you through the biggest stories of the day, helps you discover new ideas, and recommends the best in culture. Sign up for it here. It is a Monday afternoon in August, and I am on the internet watching a former cable-news anchor interview a dead teenager on Substack. This dead teenager—Joaquin Oliver, killed in the mass shooting at Marjory Stoneman Douglas High School, in Parkland, Florida—has been reanimated by generative AI, his voice and dialogue modeled on snippets of his writing and home-video footage. The animations are stiff, the model's speaking cadence is too fast, and in two instances, when it is trying to convey excitement, its pitch rises rapidly, producing a digital shriek. How many people, I wonder, had to agree that this was a good idea to get us to this moment? I feel like I'm losing my mind watching it. Jim Acosta, the former CNN personality who's conducting the interview, appears fully bought-in to the premise, adding to the surreality: He's playing it straight, even though the interactions are so bizarre. Acosta asks simple questions about Oliver's interests and how the teenager died. The chatbot, which was built with the full cooperation of Oliver's parents to advocate for gun control, responds like a press release: 'We need to create safe spaces for conversations and connections, making sure everyone feels seen.' It offers bromides such as 'More kindness and understanding can truly make a difference.' On the live chat, I watch viewers struggle to process what they are witnessing, much in the same way I am. 'Not sure how I feel about this,' one writes. 'Oh gosh, this feels so strange,' another says. Still another thinks of the family, writing, 'This must be so hard.' Someone says what I imagine we are all thinking: 'He should be here.' [Read: AI's real hallucination problem] The Acosta interview was difficult to process in the precise way that many things in this AI moment are difficult to process. I was grossed out by Acosta for 'turning a murdered child into content,' as the critic Parker Molloy put it, and angry with the tech companies that now offer a monkey's paw in the form of products that can reanimate the dead. I was alarmed when Oliver's father told Acosta during their follow-up conversation that Oliver 'is going to start having followers,' suggesting an era of murdered children as influencers. At the same time, I understood the compulsion of Oliver's parents, still processing their profound grief, to do anything in their power to preserve their son's memory and to make meaning out of senseless violence. How could I possibly judge the loss that leads Oliver's mother to talk to the chatbot for hours on end, as his father described to Acosta—what could I do with the knowledge that she loves hearing the chatbot say 'I love you, Mommy' in her dead son's voice? The interview triggered a feeling that has become exceedingly familiar over the past three years. It is the sinking feeling of a societal race toward a future that feels bloodless, hastily conceived, and shruggingly accepted. Are we really doing this? Who thought this was a good idea? In this sense, the Acosta interview is just a product of what feels like a collective delusion. This strange brew of shock, confusion, and ambivalence, I've realized, is the defining emotion of the generative-AI era. Three years into the hype, it seems that one of AI's enduring cultural impacts is to make people feel like they're losing it. During his interview with Acosta, Oliver's father noted that the family has plans to continue developing the bot. 'Any other Silicon Valley tech guy will say, 'This is just the beginning of AI,'' he said. ''This is just the beginning of what we're doing.'' Just the beginning. Perhaps you've heard that too. 'Welcome to the ChatGPT generation.' 'The Generative AI Revolution.' 'A new era for humanity,' as Mark Zuckerberg recently put it. It's the moment before the computational big bang—everything is about to change, we're told; you'll see. God may very well be in the machine. Silicon Valley has invented a new type of mind. This is a moment to rejoice—to double down. You're a fool if you're not using it at work. It is time to accelerate. How lucky we are to be alive right now! Yes, things are weird. But what do you expect? You are swimming in the primordial soup of machine cognition. There are bound to be growing pains and collateral damage. To live in such interesting times means contending with MechaHitler Grok and drinking from a fire hose of fascist-propaganda slop. It means Grandpa leaving confused Facebook comments under rendered images of Shrimp Jesus or, worse, falling for a flirty AI chatbot. This future likely requires a new social contract. But also: AI revenge porn and 'nudify' apps that use AI to undress women and children, and large language models that have devoured the total creative output of humankind. From this morass, we are told, an 'artificial general intelligence' will eventually emerge, turbo-charging the human race or, well, maybe destroying it. But look: Every boob with a T-Mobile plan will soon have more raw intelligence in their pocket than has ever existed in the world. Keep the faith. Breathlessness is the modus operandi of those who are building out this technology. The venture capitalist Marc Andreessen is quote-tweeting guys on X bleating out statements such as 'Everyone I know believes we have a few years max until the value of labor totally collapses and capital accretes to owners on a runaway loop—basically marx' worst nightmare/fantasy.' How couldn't you go a bit mad if you took them seriously? Indeed, it seems that one of the many offerings of generative AI is a kind of psychosis-as-a-service. If you are genuinely AGI-pilled—a term for those who believe that machine-born superintelligence is coming, and soon—the rational response probably involves some combination of building a bunker, quitting your job, and joining the cause. As my colleague Matteo Wong wrote after spending time with people in this cohort earlier this year, politics, the economy, and current events are essentially irrelevant to the true believers. It's hard to care about tariffs or authoritarian encroachment or getting a degree if you believe that the world as we know it is about to change forever. There are maddening effects downstream of this rhetoric. People have been involuntarily committed or had delusional breakdowns after developing relationships with chatbots. These stories have become a cottage industry in themselves, each one suggesting that a mix of obsequious models, their presentation of false information as true, and the tools' ability to mimic human conversation pushes vulnerable users to think they've developed a human relationship with a machine. Subreddits such as r/MyBoyfriendIsAI, in which people describe their relationships with chatbots, may not be representative of most users, but it's hard to browse through the testimonials and not feel that, just a few years into the generative-AI era, these tools have a powerful hold on people who may not understand what it is they're engaging with. As all of this happens, young people are experiencing a phenomenon that the writer Kyla Scanlon calls the 'End of Predictable Progress.' Broadly, the theory argues that the usual pathways to a stable economic existence are no longer reliable. 'You're thinking: These jobs that I rely on to get on the bottom rung of my career ladder are going to be taken away from me' by AI, she recently told the journalist Ezra Klein. 'I think that creates an element of fear.' The feeling of instability she describes is a hallmark of the generative-AI era. It's not at all clear yet how many entry-level jobs will be claimed by AI, but the messaging from enthusiastic CEOs and corporations certainly sounds dire. In May, Dario Amodei, the CEO of Anthropic, warned that AI could wipe out half of all entry-level white-collar jobs. In June, Salesforce CEO Marc Benioff suggested that up to 50 percent of the company's work was being done by AI. The anxiety around job loss illustrates the fuzziness of this moment. Right now, there are competing theories as to whether AI is having a meaningful effect on employment. But real and perceived impact are different things. A recent Quinnipiac poll found that, 'when it comes to their day-to-day life,' 44 percent of surveyed Americans believe that AI will do more harm than good. The survey found that Americans believe the technology will cause job loss—but many workers appeared confident in the security of their own job. Many people simply don't know what conclusions to draw about AI, but it is impossible not to be thinking about it. OpenAI CEO Sam Altman has demonstrated his own uncertainty. In a blog post titled 'The Gentle Singularity' published in June, Altman argued that 'we are past the event horizon' and are close to building digital superintelligence, and that 'in some big sense, ChatGPT is already more powerful than any human who has ever lived.' He delivered the classic rhetorical flourishes of AI boosters, arguing that 'the 2030s are likely going to be wildly different from any time that has come before.' And yet, this post also retreats ever so slightly from the dramatic rhetoric of inevitable 'revolution' that he has previously employed. 'In the most important ways, the 2030s may not be wildly different,' he wrote. 'People will still love their families, express their creativity, play games, and swim in lakes'—a cheeky nod to the endurance of our corporeal form, as a little treat. Altman is a skilled marketer, and the post might simply be a way to signal a friendlier, more palatable future for those who are a little freaked out. But a different way to read the post is to see Altman hedging slightly in the face of potential progress limitations on the technology. Earlier this month, OpenAI released GPT-5, to mixed reviews. Altman had promised 'a Ph.D.-level' intelligence on any topic. But early tests of GPT-5 revealed all kinds of anecdotal examples of sloppy answers to queries, including hallucinations, simple-arithmetic errors, and failures in basic reasoning. Some power users who'd become infatuated with previous versions of the software were angry and even bereft by the update. Altman placed particular emphasis on the product's usability and design: Paired with the 'Gentle Singularity,' GPT-5 seems like an admission that superintelligence is still just a concept. [Read: The new ChatGPT resets the AI race] And yet, the philosopher role-play continues. Not long before the launch, Altman appeared on the comedian Theo Von's popular podcast. The discussion veered into the thoughtful science-fiction territory that Altman tends to inhabit. At one point, the two had the following exchange: Sam Altman: I do guess that a lot of the world gets covered in data centers over time. Theo Von: Do you really? Altman: But I don't know, because maybe we put them in space. Like, maybe we build a big Dyson sphere around the solar system and say, 'Hey, it actually makes no sense to put these on Earth.' Von: Yeah. Altman: I wish I had, like, more concrete answers for you, but, like, we're stumbling through this. What exactly is a person, listening in their car on the way to the grocery store, to make of conversations like this? Surely, there's a cohort that finds covering the Earth or atmosphere with data centers very exciting. But what about those of us who don't? Altman and lesser personalities in the AI space often talk this way, making extreme, matter-of-fact proclamations about the future and sounding like kids playing a strategy game. This isn't a business plan; it's an idle daydream. Similarly disorienting is the fact that these visions and pontifications are driving change in the real world. Even if you personally don't believe in the hype, you are living in an economy that has reoriented itself around AI. A recent report from The Wall Street Journal estimates that Big Tech's spending on IT infrastructure in 2025 is 'acting as a sort of private-sector stimulus program,' with the 'Magnificent Seven' tech companies—Meta, Alphabet, Microsoft, Amazon, Apple, Nvidia, and Tesla—spending more than $100 billion on capital expenditures in the recent months. The flip side of such consolidated investment in one tech sector is a giant economic vulnerability that could lead to a financial crisis. This is the AI era in a nutshell. Squint one way, and you can portray it as the saving grace of the world economy. Look at it more closely, and it's a ticking time bomb lodged in the global financial system. The conversation is always polarized. Keep the faith. It's difficult to deny that generative-AI tools are transformative, insomuch as their adoption has radically altered the economy and the digital world. Social networks and the internet at large have been flooded with AI slop and synthetic text. Spotify and YouTube are filling up with AI-generated songs and videos, some of which get millions of streams. Bots are everywhere, and they have produced profoundly strange and meaningful effects on digital life. Sometimes they're racist. Many are sycophants. Other times, they summon demons. Google's AI summaries are cratering traffic and rewiring the web. In schools, ChatGPT hasn't just killed the student essay; it seems to be threatening some of the basic building blocks of human cognition. Some research has argued that chatbots are homogenizing the way people speak. In any case, they appear to have inverted the promise of the internet as an endless archive of information one can navigate for themselves. Do your own research has, in short order, become Get one canonical answer. Sometimes this is helpful: A bot artfully summarizes a complex PDF. They are, by most accounts, truly helpful coding tools. Kids use them to build helpful study guides. They're good at saving you time by churning out anemic emails. Also, a health-care chatbot made up fake body parts. The FDA has introduced a generative-AI tool to help fast-track drug and medical-device approvals—but the tool keeps making up fake studies. To scan the AI headlines is a daily exercise in trying to determine the cost that society is paying for these perceived productivity benefits. For example, with a new Google Gemini–enabled smartwatch, you can ask the bot to 'tell my spouse I'm 15 minutes late and send it in a jokey tone' instead of communicating yourself. This is followed by news of a study suggesting that ChatGPT power users might be accumulating a 'cognitive debt' from using the tool. In recent months, I've felt unmoored by all of this: by a technology that I find useful in certain contexts being treated as a portal to sentience; by a billionaire confidently declaring that he is close to making breakthroughs in physics by conversing with a chatbot; by a 'get that bag' culture that seems to have accepted these tools without much consideration as to the repercussions; by the discourse. I hear the chatter everywhere—a guy selling produce at the farmers' market makes a half-hearted joke that AI can't grow blueberries; a woman at the airport tells her friend that she asked ChatGPT for makeup recommendations. Most of these conversations are poorly informed, conducted by people who have been bombarded for years now by hype but who have also watched as some of these tools have become ingrained in their life or in the life of people they know. They're not quite excited or jaded, but almost all of them seem resigned to dealing with the tools as part of their future. Remember—this is just the beginning … right? This is the language that the technology's builders and backers have given us, which means that discussions that situate the technology in the future are being had on their terms. This is a mistake, and it is perhaps the reason so many people feel adrift. Lately, I've been preoccupied with a different question: What if generative AI isn't God in the machine or vaporware? What if it's just good enough, useful to many without being revolutionary? Right now, the models don't think—they predict and arrange tokens of language to provide plausible responses to queries. There is little compelling evidence that they will evolve without some kind of quantum research leap. What if they never stop hallucinating and never develop the kind of creative ingenuity that powers actual human intelligence? The models being good enough doesn't mean that the industry collapses overnight or that the technology is useless (though it could). The technology may still do an excellent job of making our educational system irrelevant, leaving a generation reliant on getting answers from a chatbot instead of thinking for themselves, without the promised advantage of a sentient bot that invents cancer cures. [Read: AI executives promise cancer cures. Here's the reality.] Good enough has been keeping me up at night. Because good enough would likely mean that not enough people recognize what's really being built—and what's being sacrificed—until it's too late. What if the real doomer scenario is that we pollute the internet and the planet, reorient our economy and leverage ourselves, outsource big chunks of our minds, realign our geopolitics and culture, and fight endlessly over a technology that never comes close to delivering on its grandest promises? What if we spend so much time waiting and arguing that we fail to marshal our energy toward addressing the problems that exist here and now? That would be a tragedy—the product of a mass delusion. What scares me the most about this scenario is that it's the only one that doesn't sound all that insane. Article originally published at The Atlantic

Ethics complaints filed against Sewall's Point officials following clerk's firing
Ethics complaints filed against Sewall's Point officials following clerk's firing

Yahoo

time4 hours ago

  • Yahoo

Ethics complaints filed against Sewall's Point officials following clerk's firing

SEWALL'S POINT — Two public officials here have been slapped with ethics complaints in the wake of the town clerk's dismissal. Former Town Clerk April Stoncius filed the complaints against Finance and Human Resource Director Maria Pierce and Town Manager Bob Daniels June 28 and July 11, respectively. Copies of the complaints were obtained by TCPalm. Daniels fired Stoncius for insubordination, Stoncius said. Her last day as town clerk was April 28. "The documents speak for themselves and reflect a clear breach of ethics and integrity, as well as a violation of public trust," Stoncius said by email. The Florida Commission on Ethics doesn't confirm nor deny whether complaints have been filed, commission spokesperson Lynn Blais said. Both Daniels and Pierce declined Aug. 19 to comment for this story. But Town Attorney Glen Torcivia, who teaches ethics courses to such groups as the Treasure Coast Regional League of Cities, said he sees no merit to either complaint. The complaint against Pierce Pierce allegedly violated the town's public records request policy, the complaint shows, by responding to a request from the media without notifying the town clerk. Pierce also allegedly changed the format of a commission meeting agenda without consulting with the town clerk, "resulting in the first unapproved agenda in the town's history," Stoncius wrote in the complaint. Because the agenda went unapproved, none of the business at the meeting was binding, Stoncius said in the complaint. Additionally, someone altered commission meeting minutes, which Stoncius was told to send to Pierce before sending them to town commissioners. "Any altering of what occurred at a public meeting for one's own benefit is unethical at best," Stoncius wrote in the complaint. What was altered was a comment Pierce made to the commission, the complaint says. The complaint against Daniels Daniels refused to release a report on septic-to-sewer conversions in response to a public records request Oct. 24, 2022, Stoncius' complaint says. As of July 11, the report had yet to be released, according to the complaint. Torcivia, however, said the report was simply mentioned in a separate report by Town Engineer Joe Capra of Captec Engineering. The requested report also never was prepared for the town, Torcivia said. Daniels also allegedly interfered with and failed to provide an answer to the town clerk regarding a public records request filed October 2024 by Frank Tidikis, so the clerk could process it according to state law. A month later Tidikis was elected to the Town Commission. Stoncius — following no response from Daniels or the town engineer — took it upon herself to fulfill the request, the complaint says. "After fulfilling now-Commissioner Tidikis' public records request, I experienced 'heat' from Mr. Daniels for providing the correct information," Stoncius wrote in the complaint. More on crime: Florida Turnpike crash: What to know about driver charged and three people who died More on Stuart: Southern Living magazine names this Florida town as 'best coastal small town in the U.S.' Lastly, Stoncius alleges, Daniels never responded to a Jan. 23, public records request regarding maps of the town's stormwater drainage system. In that case, the requestor should have come forward and complained, Torcivia said. "I don't think that claim has any merit at all," he said. "I have been harassed and treated unfairly for providing the truth in the process ofperforming my sworn position," Stoncius says in the complaint, "in an effort to provide transparency to the public. There is a lack of ethics and integrity under Mr. Daniels' management." Keith Burbank is TCPalm's watchdog reporter covering Martin County. He can be reached at and at 720-288-6882. This article originally appeared on Treasure Coast Newspapers: Ethics complaints lodged against two Sewall's Point officials Solve the daily Crossword

Woman killed in apartment fire in Pittsburgh's Brighton Heights neighborhood
Woman killed in apartment fire in Pittsburgh's Brighton Heights neighborhood

CBS News

time12 hours ago

  • CBS News

Woman killed in apartment fire in Pittsburgh's Brighton Heights neighborhood

A woman is dead after a fire at an apartment in Pittsburgh's Brighton Heights neighborhood on Tuesday. Pittsburgh Public Safety said firefighters responded to a fire at Riverview Manor on Letort Street at 8:15 p.m. Officials say a woman was pulled from a unit in cardiac arrest and was pronounced dead at the scene. A man also suffered a wrist laceration, Public Safety said. Emergency management officials have been called in to work with building management and help find temporary accommodations for any displaced residents. There's been no word on the cause, but Public Safety said fire investigators are looking into it.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store