
Paedophile had 5,000 images of children and AI generated images of child abuse
Paedophile had 5,000 images of children and AI generated images of child abuse
Jack Gillard, 25, was found in possession of a category A video which depicted a compilation of children aged between nine and 15-years-old being abused
Jack Gillard, 25, was found in possession of more than 5,000 indecent images of children
(Image: South Wales Police )
A paedophile was found in possession of more than 5,000 indecent images of children as young as six-years-old as well as AI generated images of child abuse. His family blamed his offending on "fixated behaviour".
Jack Gillard, 25, was visited by police at his home after intelligence had been received about the downloading of child abuse images. A raid was carried out on July 26, 2023, and a number of devices were seized including the defendant's mobile phone.
A sentencing hearing at Merthyr Tydfil Crown Court on Friday heard Gillard was arrested and made no comment. For the latest court reports sign up to our crime newsletter.
He was taken to Merthyr Tydfil police station and again made no comment.
Prosecutor Hashim Salmman said a digital forensic examination was conducted on the seized devices and more than 5,000 category A, B and C images were discovered on the defendant's phone.
There was also prohibited material generated by AI which depicted young girls being raped by adult men.
Article continues below
One category A video was accessible, and was described as a compilation video including children aged between nine and 15-years-old involved in various types of sexual activity with adult males.
The youngest child depicted in the images was just six-years-old.
A number of screenshots were found of websites and forums including a collage of images.
There were a total of 1,476 category A images, 370 category B images and 3,145 category C images.
Gillard, of Trealaw Road, Trealaw, Tonypandy, pleaded guilty to three counts of making indecent images of children. The court heard he has no previous convictions.
In mitigation, it was said the defendant's behaviour would "never happen again". It was also said his family have concerns about "fixating behaviour, but he has been working hard on reducing his cannabis consumption.
Sentencing, the Recorder of Cardiff, Judge Tracey Lloyd-Clarke, said the defendant was a risk to children due to his "sexual interest in pre pubescent girls".
Gillard was sentenced to one year and 10 months imprisonment.
He was also made subject to a Sexual Harm Prevention Order and sex offender notification requirements for 10 years.
Article continues below

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Daily Mail
2 hours ago
- Daily Mail
Lawyers warned to stop using ChatGPT to argue lawsuits after AI programs 'made up fictitious cases'
Lawyers in England and Wales have been warned they could face 'severe sanctions' including potential criminal prosecution if they present false material generated by AI in court. The ruling, by one of Britain's most senior judges, comes on the back of a string of cases in which which artificially intelligence software has produced fictitious legal cases and completely invented quotes. The first case saw AI fabricate 'inaccurate and fictitious' material in a lawsuit brought against two banks, The New York Times reported. Meanwhile, the second involved a lawyer for a man suing his local council who was unable to explain the origin of the nonexistent precedents in his legal argument. While large language models (LLMs) like OpenAI 's ChatGPT and Google 's Gemini are capable of producing long accurate-sounding texts, they are technically only focused on producing a 'statistically plausible' reply. The programs are also prone to what researchers call 'hallucinations' - outputs that are misleading or lack any factual basis. AI Agent and Assistance platform Vectera has monitored the accuracy of AI chatbots since 2023 and found that the top programs hallucinate between 0.7 per cent and 2.2 per cent of the time - with others dramatically higher. However, those figures become astronomically higher when the chatbots are prompted to produce longer texts from scratch, with market leader OpenAI recently acknowledging that its flagship ChatGPT system hallucinates between 51 per cent and 79 per cent of the time if asked open-ended questions. While large language models (LLMs) like OpenAI's ChatGPT and Google's Gemini are capable of producing long accurate-sounding texts, they are technically only focused on producing a 'statistically plausible' reply - which can lead to them 'hallucinating' false information Dame Victoria Sharp, president of the King's Bench Division of the High Court, and Justice Jeremy Johnson KC, authored the new ruling. In it they say: 'The referrals arise out of the actual or suspected use by lawyers of generative artificial intelligence tools to produce written legal arguments or witness statements which are not then checked, so that false information (typically a fake citation or quotation) is put before the court. 'The facts of these cases raise concerns about the competence and conduct of the individual lawyers who have been referred to this court. 'They raise broader areas of concern however as to the adequacy of the training, supervision and regulation of those who practice before the courts, and as to the practical steps taken by those with responsibilities in those areas to ensure that lawyers who conduct litigation understand and comply with their professional and ethical responsibilities and their duties to the court.' The pair argued that existing guidance around AI was 'insufficient to address the misuse of artificial intelligence'. Judge Sharp wrote: 'There are serious implications for the administration of justice and public confidence in the justice system if artificial intelligence is misused,' While acknowledging that AI remained a 'powerful technology' with legitimate use cases, she nevertheless reiterated that the technology brought 'risks as well as opportunities.' In the first case cited in the judgment, a British man sought millions in damages from two banks. The court discovered that 18 out of 45 citations included in the legal arguments featured past cases that simply did not exist. Even in instances in which the cases did exist, often the quotations were inaccurate or did not support the legal argument being presented. The second case, which dates to May 2023, involved a man who was turned down for emergency accommodation from the local authority and ultimately became homeless. His legal team cited five past cases, which the opposing lawyers discovered simply did not exist - tipped off by the fact by the US spellings and formulaic prose style. Rapid improvements in AI systems means its use is becoming a global issue in the field of law, as the judicial sector figures out how to incorporate artificial intelligence into what is frequently a very traditional, rules-bound work environment. Earlier this year a New York lawyer faced disciplinary proceedings after being caught using ChatGPT for research and citing a none-existent case in a medical malpractice lawsuit. Attorney Jae Lee was referred to the grievance panel of the 2nd U.S. Circuit Court of Appeals in February 2025 after she cited a fabricated case about a Queens doctor botching an abortion in an appeal to revive her client's lawsuit. The case did not exist and had been conjured up by OpenAI's ChatGPT and the case was dismissed. The court ordered Lee to submit a copy of the cited decision after it was not able to find the case. She responded that she was 'unable to furnish a copy of the decision.' Lee said she had included a case 'suggested' by ChatGPT but that there was 'no bad faith, willfulness, or prejudice towards the opposing party or the judicial system' in doing so. The conduct 'falls well below the basic obligations of counsel,' a three-judge panel for the Manhattan-based appeals court wrote. In June two New York lawyers were fined $5,000 after they relied on fake research created by ChatGPT for a submission in an injury claim against Avianca airline. Judge Kevin Castel said attorneys Steven Schwartz and Peter LoDuca acted in bad faith by using the AI bot's submissions - some of which contained 'gibberish' - even after judicial orders questioned their authenticity.


Scottish Sun
5 hours ago
- Scottish Sun
Man, 50, dies at UK beauty spot after getting into trouble swimming in water
Coastguard crews, police and paramedics were called to the scene BEACH TRAGEDY Man, 50, dies at UK beauty spot after getting into trouble swimming in water A MAN has tragically died after getting into difficulty while swimming near a UK beauty spot. The alarm was raised at Tor Bay, near Penmaen village on the Gower Peninsula, in South Wales, this afternoon. Coastguard crews, police and paramedics were called to the scene. A 50-year-old from Sandfields, Swansea, was pulled from the water at around 1pm, said South Wales Police. He was pronounced dead at the scene by paramedics at around 4pm. SWP said the man's family has been informed. 1 A man has died after getting into difficulty while swimming near a beauty spot in South Wales Credit: Supplied More to follow... For the latest news on this story keep checking back at The Sun Online is your go-to destination for the best celebrity news, real-life stories, jaw-dropping pictures and must-see video. Like us on Facebook at and follow us from our main Twitter account at @TheSun.


The Sun
5 hours ago
- The Sun
Man, 50, dies at UK beauty spot after getting into trouble swimming in water
A MAN has tragically died after getting into difficulty while swimming near a UK beauty spot. The alarm was raised at Tor Bay, near Penmaen village on the Gower Peninsula, in South Wales, this afternoon. Coastguard crews, police and paramedics were called to the scene. A 50-year-old from Sandfields, Swansea, was pulled from the water at around 1pm, said South Wales Police. He was pronounced dead at the scene by paramedics at around 4pm. SWP said the man's family has been informed. 1 is your go-to destination for the best celebrity news, real-life stories, jaw-dropping pictures and must-see video. Like us on Facebook at and follow us from our main Twitter account at @TheSun.