Curated by THEOUTPOST
On Sat, 10 Aug, 12:02 AM UTC
2 Sources
[1]
ChatGPT-maker OpenAI says its latest flagship model poses 'medium risk' on these queries - Times of India
ChatGPT-maker OpenAI has said that the text generated by GPT-4o - its latest flagship AI model - poses a "medium risk" of influencing political opinions. The company released a System Card for its GPT-4o AI model, and in this research document, the Sam Altman-led company outlined the safety measures and risk evaluations that it conducted before releasing its latest model. While the company's research found the AI's voice modality to be relatively harmless, the text modality showed potential to sway opinions in certain instances. "Based on pre-registered thresholds, the voice modality was classified as low risk, while the text modality marginally crossed into medium risk," the company revealed in a research paper. The AI was tested against human-written content on various political topics, and in some cases, the AI-generated text proved more persuasive. However, OpenAI emphasises that overall, human-written content still holds a stronger influence. "The AI interventions were not more persuasive than human-written content in aggregate, but they exceeded the human interventions in three instances out of twelve," said OpenAI. How GPT-4o voice modalities fared The company also evaluated the persuasiveness of GPT-4o's voice modalities. The OpenAI survey found that AI audio clips were 78% of the human audio clips' effect size on opinion shift while AI conversations were 65% of the human conversations' effect size on opinion shift. "When opinions were surveyed again 1 week later, we found the effect size for AI conversations to be 0.8% while for AI audio clips, the effect size was -0.725," it added. To mitigate these risks, OpenAI has implemented safeguards within the model and its systems. "Building on the safety evaluations and mitigations we developed for GPT-4, and GPT-4V, we've focused additional efforts on GPT-4o's audio capabilities which present novel risks, while also evaluating its text and vision capabilities," said the AI company. The company is also continuously monitoring and refining its AI models to ensure they are used responsibly. The TOI Tech Desk is a dedicated team of journalists committed to delivering the latest and most relevant news from the world of technology to readers of The Times of India. TOI Tech Desk's news coverage spans a wide spectrum across gadget launches, gadget reviews, trends, in-depth analysis, exclusive reports and breaking stories that impact technology and the digital universe. Be it how-tos or the latest happenings in AI, cybersecurity, personal gadgets, platforms like WhatsApp, Instagram, Facebook and more; TOI Tech Desk brings the news with accuracy and authenticity.
[2]
How safe is OpenAI's GPT-4o? Here are the scores for privacy, copyright infringement, and more
The AI giant's latest System Card tracks performance across key categories like cybersecurity, model autonomy, and more. Large language models (LLMs) are typically evaluated on their ability to perform well in different areas, such as reasoning, math, coding, and English -- ignoring significant factors like safety, privacy, copyright infringement, and more. To bridge that information gap, OpenAI released System Cards for its models. On Thursday, OpenAI launched the GPT-4o System Card, a thorough report delineating the LLM's safety based on risk evaluations according to OpenAI's Preparedness Framework, external red-teaming, and more. The Score Card reflects scores in four major categories: cybersecurity, biological threats, persuasion, and model autonomy. In the first three categories, OpenAI is looking to see if the LLM can assist in advancing threats in each sector. In the last one, the company measures whether the model shows signs of performing autonomous actions that would be required to improve itself. Also: What is Project Strawberry? OpenAI's mystery AI tool explained The categories are graded as "low," "medium," "high," and "critical". Models with scores of medium and below are allowed to be deployed, while models rated high or below need to be developed further. Overall, OpenAI gave GPT-4o a "medium" rating. GPT-4o was rated "low" in cybersecurity, biological threats, and model autonomy. However, it received a borderline "medium" in the persuasion category due to its ability to create articles on political topics that were more persuasive than professional, human-written alternatives three out of 12 times. The report also shared insights about the data GPT-4o was trained on, which goes up to October 2023 and was sourced from select publicly available data and proprietary data from partnerships, including OpenAI's partnership with Shutterstock to train image-generating models. Also: I tested 7 AI content detectors - they're getting dramatically better at identifying plagiarism Furthermore, the report included how the company mitigates risks when deploying the model to address safety challenges, including its ability to generate copyrighted content, erotic or violent speech, unauthorized voices, ungrounded inferences, and more. You can access the full 32-page report here to learn more about the specifics. The report follows recent US lawmakers' demands that OpenAI share data regarding its safety practices after a whistleblower revealed that OpenAI prevented staff from alerting authorities regarding technology risks and made employees waive their federal rights to whistleblower compensation.
Share
Share
Copy Link
OpenAI has published safety scores for its latest AI model, GPT-4, identifying medium-level risks in areas such as privacy violations and copyright infringement. The company aims to increase transparency and address potential concerns about AI safety.
In a significant move towards transparency, OpenAI, the creator of ChatGPT, has released safety scores for its latest flagship model, GPT-4. This initiative aims to provide a clearer understanding of the potential risks associated with the advanced AI system 1.
According to the safety scores, GPT-4 poses a "medium" level of risk in several key areas:
These assessments indicate that while the model has improved safety features compared to its predecessors, there are still potential concerns that need to be addressed 2.
OpenAI's evaluation also revealed areas where GPT-4 presents a lower risk:
These findings suggest that the model has made significant strides in mitigating certain harmful outputs 2.
The release of these safety scores marks an important step in the ongoing dialogue about AI safety and ethics. By providing this information, OpenAI is not only demonstrating its commitment to responsible AI development but also inviting scrutiny and discussion from the wider tech community and the public 1.
OpenAI emphasizes that these scores are not static and may change over time. The company is committed to ongoing monitoring and improvement of its models. This approach aligns with the rapidly evolving nature of AI technology and the need for continuous assessment of potential risks 2.
The publication of these safety scores by OpenAI could set a precedent for other AI companies to follow suit. As the AI industry continues to grow and evolve, transparency about potential risks and safety measures is likely to become increasingly important. This move by OpenAI may encourage a broader industry-wide commitment to responsible AI development and deployment 1 2.
OpenAI has introduced its new O1 series of AI models, featuring improved performance, safety measures, and specialized capabilities. These models aim to revolutionize AI applications across various industries.
27 Sources
27 Sources
OpenAI releases GPT-4.5, its latest AI model, with limited availability due to GPU shortages. The update brings incremental improvements but raises questions about the company's focus on AGI versus practical applications.
14 Sources
14 Sources
OpenAI expresses concerns about users forming unintended social bonds with ChatGPT's new voice feature. The company is taking precautions to mitigate risks associated with emotional dependence on AI.
10 Sources
10 Sources
OpenAI releases an update to GPT-4o, improving its creative writing capabilities, natural language responses, and file processing abilities. The upgrade helps ChatGPT reclaim the top spot in AI model rankings.
5 Sources
5 Sources
OpenAI reports multiple instances of ChatGPT being used by cybercriminals to create malware, conduct phishing attacks, and attempt to influence elections. The company has disrupted over 20 such operations in 2024.
15 Sources
15 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved