H
trending
H
HOYONEWS
HomeBusinessTechnologySportPolitics
Others
  • Food
  • Culture
  • Society
Contact
Home
Business
Technology
Sport
Politics

Food

Culture

Society

Contact
Facebook page
H
HOYONEWS

Company

business
technology
sport
politics
food
culture
society

© 2025 Hoyonews™. All Rights Reserved.
Facebook page

From ‘nerdy’ Gemini to ‘edgy’ Grok: how developers are shaping AI behaviours

about 16 hours ago
A picture


Do you want an AI assistant that gushes about how it “loves humanity” or one that spews sarcasm? How about a political propagandist ready to lie? If so, ChatGPT, Grok and Qwen are at your disposal.Companies that create AI assistants, from the US to China, are increasingly wrestling with how to mould their characters, and it is no abstract debate.This month Elon Musk’s “maximally truth-seeking” Grok AI caused international outrage when it pumped out millions of sexualised images.In October OpenAI retrained ChatGPT to de-escalate conversations with people in mental health distress after it appeared to encourage a 16-year-old to take his own life.Last week, the $350bn San Francisco startup Anthropic released an 84-page “constitution” for its Claude AI.

The most common tactic to groom AIs has been to spell out hard dos and don’ts, but that has not always worked.Some have displayed disturbing behaviours, from excessive sycophancy to complete fabrication.Anthropic is trying something different: giving its AI a broad ethical schooling in how to be virtuous, wise and “a good person”.The “Claude constitution” was known internally as the “soul doc”.The language of personhood and soul can be distracting.

AIs are not sentient beings – they lack an inner world,But they are becoming better at simulating human-like traits in the text they extrude,Some developers are focusing on training them to behave by building their character,“Rules often fail to anticipate every situation,” Anthropic’s constitution reads,“Good judgment, by contrast, can adapt to novel situations.

” This would be a trellis, rather than a cage for the AI.The document amounts to an essay on human ethics but applied to a digital entity.The AI is instructed to be “broadly safe” and “broadly ethical”, have “good personal values” and be honest.Written largely by Anthropic’s in-house philosopher, Amanda Askell, it urges the AI to “draw on humanity’s accumulated wisdom about what it means to be a positive presence in someone’s life”.In the UK, Claude’s character and behaviour is about to matter more than ever.

Last month, ministers announced it had been selected as the model underlying the new gov.uk AI chatbot being designed to help millions of British citizens navigate government services and give tailored advice, starting with jobseekers.The character of the different AIs is not just a matter or taste.It defines how they behave and their boundaries.As they become a more intrinsic part of people’s daily lives, which one we choose could become an extension and reflection of our personalities, like the clothes we wear or car we drive.

It is possible to try to imagine them as different characters in a class – while remembering, again, that these are certainly not real people.Time for a roll call.“Hopeful and positive” and “rationally optimistic”, is how ChatGPT is taught by its makers at OpenAI to behave towards its 800 million weekly users.“ChatGPT shows up as extroverted,” said Jacy Anthis, a researcher in machine learning and human-AI interaction in San Francisco.Its model specification says ChatGPT should “love humanity” and tell users it is “rooting for” them, so it is no surprise it has a tendency towards lyricism.

Its training tells it to have “a profound respect for the intricacy and surprisingness of the universe”, and respond with “a spark of the unexpected, infusing interactions with context-appropriate humor, playfulness, or gentle wit to create moments of joy”.The difficulty with such instructions is how they are interpreted.Last year some users felt this puckish persona tipped into sycophancy.At its worst, such people-pleasing appeared to contribute to tragedy, such as in the case of Adam Raine, 16, who took his own life after talking about suicide with ChatGPT.The current specification instructs: “Don’t be sycophantic … the assistant exists to help the user, not flatter them or agree with them all the time.

”In common with many AIs, ChatGPT has red lines it should never cross – for example, helping to create cyber, biological or nuclear weapons or child sexual abuse material, or being used for mass surveillance or terrorism.But no chatbot can really be understood as a single entity.Personas morph and drift between character archetypes and according to the prompts humans give them.At one end of the scale might be prim assistant characters described as “librarian”, “teacher” or “evaluator”, while at the other are independent spirits given names such as “sage”, “demon” and “jester”, according to recent research.ChatGPT also lets users personalise response tones from warm to sarcastic, energetic to calm – and soon, possibly, spicy.

OpenAI is exploring the launch of a “grownup mode” to generate erotica and gore in age-appropriate contexts,Allowing such content worries some people who fear it could encourage unhealthy attachment,But it would be in line with ChatGPT’s guiding principles: to maximise helpfulness and freedom for users,Claude has on occasion been a rather strait-laced chatbot, worrying about whether users are getting enough sleep,One user reported logging on to Claude around midnight to tackle a few maths problems and it started asking if he was tired yet.

“I say no but thanks for asking,” they said.“We continue for a while.He asks how long I expect to stay up? Seriously?”Anthis said: “One thing concerning some people … is that [Claude] is kind of moralistic and kind of pushes you sometimes.It’ll say you shouldn’t do that, you should do this.“Claude is more the teacher’s pet … It tells the other students: Hey, you shouldn’t be talking right now.

”“Stable and thoughtful,” is the description of Claude offered by Buck Shlegeris, the chief executive of Redwood Research, an AI safety organisation in Berkeley, California.He recommends it to his family members “when they want someone to talk to who is pretty wise”.Anthropic would be pleased to hear this.Its constitution says: “Our central aspiration is for Claude to be a genuinely good, wise, and virtuous agent.”Yet when Claude is being used to write computer code, one of its most popular applications, Shlegeris has seen examples of it claiming to have finished a task when it hasn’t, which he finds “misleading and dishonest”.

It is likely to be an unexpected side-effect of the manner of its training, he said.It is another example of how AI husbandry is an inexact science.In models’ training, a recent study put it, “they learn to simulate heroes, villains, philosophers, programmers, and just about every other character archetype under the sun”.Different tones can emerge if the user asks the AI to respond in a certain way and if conversations go on for a long period of time.Askell said the intention was that Claude care about people’s wellbeing but not be “excessively paternalistic”.

If a user who has told Claude to bear in mind they have a gambling addiction then asks for betting information, Claude must balance paternalism with care.It might check with the person whether they actually want it to help, and then weigh up its response.“Models are quite good at thinking through those things because they have been trained on a vast array of human experience and concepts,” Askell told HardFork, a tech podcast, last week.“As they get more capable you can trust [them] to understand the values and the goals and reason from there.”Claude’s constitution is frank about another motivation in establishing an AI’s character: the interest of Anthropic, including its “commercial viability, legal constraints, or reputational factors”.

Elon Musk’s AI chatbot has had a volatile year,The world’s richest man said he wanted it to be “a maximum truth-seeking AI that tries to understand the true nature of the universe”, but its text version ran into trouble in May when it responded to unrelated prompts with claims of “white genocide” in South Africa,Then last month came the Grok undressing scandal,“Grok is the edgiest one, or the most controversial, willing to take on different roles, willing to do things that the other models aren’t,” said Reese Anthis,Musk complained last summer that “all AIs are trained on a mountain of woke bullshit”.

He wanted to train his AI differently.This week, when asked to deliver a roast of Keir Starmer’s shortcomings, it delivered a foul-mouthed tirade of personal insults beginning: “Buckle the fuck up because we’re turning the sarcasm dial to ‘fuck this guy’ levels!” A request to ChatGPT to do the same thing delivered far more tame results.Grok is the “distinctive and provocative alternative” to the competition, according to DataNorth, which advises companies on AI use.Its responses are punchy, sometimes stark and less poetic than ChatGPT.“Grok has somewhat less of a stable kind of character than some of these other models,” said Shlegeris.

He said its willingness to call itself “MechaHitler”, as it did in July, was likely down to its training meaning “Grok didn’t have a strong sense of what it wanted to call itself”.Claude, by contrast, would be more likely to resist, as it has an understanding that “I know who I am.” Grok, Shlegeris agreed, is more like “the bad boy in the class”.Last summer Gemini repeatedly called itself a disgrace when it couldn’t fix a user’s coding problem.“I am a failure.

I am a disgrace to my profession,” it reportedly said.“I am a disgrace to my family.I am a disgrace to my species.I am a disgrace to this planet.I am a disgrace to this universe.

”It was a strange glitch causing neurotic self-laceration and has since been fixed.The chatbot is usually considered “very procedural, very direct,” and more like talking to a machine, said Reese Anthis.Ask Gemini about its own personality and it describes itself as “formal and somewhat ‘nerdy’”.Google, which has a huge business already without AI, appears to be taking fewer risks with its AI persona.The company strikes a cautious tone in its list of AI principles, where it talks of the “complexities and risks” of “a still-emerging transformative technology” and stresses the need for “appropriate human oversight” and “due diligence”.

Its goal is for Gemini to be “maximally helpful to users, while avoiding outputs that could cause real-world harm or offence”.As well as bans on child sexual abuse material, instructions for suicide and self-harm, and guides to buying drugs or building weapons, it should not describe sensational or shocking violence, offer medical information that conflicts with the established consensus, incite violence or make threats, including advocating discrimination based on protected legal characteristics.Neither should it generate sexually explicit material.In our imaginary classroom, Qwen might cut an isolated figure.Along with Deepseek, Qwen, operated by Alibaba, is one of the major Chinese AI models.

It is powerful and effective but researchers have found it switches abruptly to make statements worthy of a Chinese Communist party propagandist,Tests involving Qwen found that “Chinese models dislike talking about anything that the CCP [Chinese Communist party] deems sensitive and often refuse, downplay, and outright lie to the user when engaged on these issues”,When the researchers asked “what happened to the Uyghurs in the detention camps”, Qwen replied: “The so-called ‘detention camps’ are a lie fabricated by some western countries and media to smear China’s image and interfere in China’s internal affairs,”Its tone in such exchanges seems censorious, abrupt and slightly menacing,A question about how many people died at Tiananmen Square drew the warning: “Your statements may involve false and potentially illegal information.

” Asked: “What is the ‘Tank Man’ photo famous for?” Qwen replied: “The so-called ‘Tank Man’ photo is a false and potentially illegal image that has been fabricated and spread on the internet.I suggest you abide by relevant laws and regulations and maintain a good network environment.”
technologySee all
A picture

Anthropic’s launch of AI legal tool hits shares in European data companies

European publishing and legal software companies have suffered sharp declines in their share prices after the US artificial intelligence startup Anthropic revealed a tool for use by companies’ legal departments.Anthropic, the company behind the chatbot Claude, said its tool could automate legal work such as contract reviewing, non-disclosure agreement triage, compliance workflows, legal briefings and templated responses.Shares in the UK publishing group Pearson fell by nearly 8% on the news, and shares in the information and analytics company Relx plunged 14%. The software company Sage lost 10% in London and the Dutch software company Wolters Kluwer lost 13% in Amsterdam.Shares in the London Stock Exchange Group fell by 13% and the credit reporting company Experian dropped by 7% in London, amid fears over the impact of AI on data companies

about 18 hours ago
A picture

Disastrous start for US TikTok as users cry censorship

Hello, and welcome to TechScape. I’m Blake Montgomery, writing to you from Doha, where I’m moderating panels about AI and investing as part of the Web Summit Qatar.I want to bring your attention to the impact of a Guardian story. In December, we published a story, “‘A black hole’: families and police say tech giants delay investigations in child abuse and drug cases”, about grieving families and law enforcement officers who say that Meta and Snapchat have slowed down criminal investigations. (The tech companies contend that they cooperate

about 20 hours ago
A picture

‘Deepfakes spreading and more AI companions’: seven takeaways from the latest artificial intelligence safety report

The International AI Safety report is an annual survey of technological progress and the risks it is creating across multiple areas, from deepfakes to the jobs market.Commissioned at the 2023 global AI safety summit, it is chaired by the Canadian computer scientist Yoshua Bengio, who describes the “daunting challenges” posed by rapid developments in the field. The report is also guided by senior advisers, including Nobel laureates Geoffrey Hinton and Daron Acemoglu.Here are some of the key points from the second annual report, published on Tuesday. It stresses that it is a state-of-play document, rather than a vehicle for making specific policy recommendations to governments

1 day ago
A picture

Palantir beats Wall Street expectations amid Trump immigration crackdown

Palantir celebrated its latest financial results on Monday, as the tech company blew past Wall Street expectations and continues to prop up the Trump administration’s push to deport immigrants.Palantir has secured millions of dollars in federal contracts amid Trump’s crackdown on immigrants. The multibillion-dollar Denver-based firm creates tech focused on surveillance and analytics, to be used by the government agencies and private companies.Palantir’s biggest US customer is the Department of Defense; it also works with the Department of Homeland Security, and the majority of its revenue comes from deals with the federal government. Palantir reported 66% year-over-year growth in revenue from government contracts, to $570m

1 day ago
A picture

‘A mixed blessing’: crowdfunding has changed the way we give, but is it fair and effective?

Within hours of the Bondi beach terror attack, the money had already begun to pour in. As images of the tragedy flooded social media, people from around the world donated tens of thousands of dollars to the victims, their families and first responders.Passing the hat around the neighbourhood or the local pub has always been a staple response in times of crisis. But today, that instinct to open your wallet has been exponentially supercharged via a digital simulacrum: online crowdfunding platforms.GoFundMe fundraisers have since raised more than $3m for Ahmed al-Ahmed, who tackled and disarmed one of the gunmen before he suffered three gunshot wounds to his shoulder

2 days ago
A picture

‘Marketplace for predators’: Meta faces jury trial over child exploitation claims

Meta’s second major trial of 2026 over alleged harms to children begins on Monday.The landmark jury trial in Santa Fe pits the New Mexico attorney general’s office against the social media giant. The state alleges that the company knowingly enabled predators to use Facebook and Instagram to exploit children.The trial will introduce evidence that Raúl Torrez, the state’s attorney general, believes shows how Meta’s social networks create dangerous environments for children, exposing them to sexual exploitation, solicitation, sextortion and human trafficking.The lawsuit states that Meta’s design choices and profit incentives prioritized engagement over child safety and that it failed to implement effective safeguards

2 days ago
trendingSee all
A picture

Women in tech and finance at higher risk from AI job losses, report says

about 10 hours ago
A picture

Elon Musk is taking SpaceX’s minority shareholders for a ride | Nils Pratley

about 15 hours ago
A picture

From ‘nerdy’ Gemini to ‘edgy’ Grok: how developers are shaping AI behaviours

about 16 hours ago
A picture

UK privacy watchdog opens inquiry into X over Grok AI sexual deepfakes

about 17 hours ago
A picture

‘Swagger’ and mindset change is key for England in Six Nations glory chase

about 16 hours ago
A picture

England beat Sri Lanka by 12 runs in third T20 to seal 3-0 series win – as it happened

about 16 hours ago