2 Sources
[1]
ChatGPT-maker OpenAI says its latest flagship model poses 'medium risk' on these queries - Times of India
ChatGPT-maker OpenAI has said that the text generated by GPT-4o - its latest flagship AI model - poses a "medium risk" of influencing political opinions. The company released a System Card for its GPT-4o AI model, and in this research document, the Sam Altman-led company outlined the safety measures and risk evaluations that it conducted before releasing its latest model. While the company's research found the AI's voice modality to be relatively harmless, the text modality showed potential to sway opinions in certain instances. "Based on pre-registered thresholds, the voice modality was classified as low risk, while the text modality marginally crossed into medium risk," the company revealed in a research paper. The AI was tested against human-written content on various political topics, and in some cases, the AI-generated text proved more persuasive. However, OpenAI emphasises that overall, human-written content still holds a stronger influence. "The AI interventions were not more persuasive than human-written content in aggregate, but they exceeded the human interventions in three instances out of twelve," said OpenAI. How GPT-4o voice modalities fared The company also evaluated the persuasiveness of GPT-4o's voice modalities. The OpenAI survey found that AI audio clips were 78% of the human audio clips' effect size on opinion shift while AI conversations were 65% of the human conversations' effect size on opinion shift. "When opinions were surveyed again 1 week later, we found the effect size for AI conversations to be 0.8% while for AI audio clips, the effect size was -0.725," it added. To mitigate these risks, OpenAI has implemented safeguards within the model and its systems. "Building on the safety evaluations and mitigations we developed for GPT-4, and GPT-4V, we've focused additional efforts on GPT-4o's audio capabilities which present novel risks, while also evaluating its text and vision capabilities," said the AI company. The company is also continuously monitoring and refining its AI models to ensure they are used responsibly. The TOI Tech Desk is a dedicated team of journalists committed to delivering the latest and most relevant news from the world of technology to readers of The Times of India. TOI Tech Desk's news coverage spans a wide spectrum across gadget launches, gadget reviews, trends, in-depth analysis, exclusive reports and breaking stories that impact technology and the digital universe. Be it how-tos or the latest happenings in AI, cybersecurity, personal gadgets, platforms like WhatsApp, Instagram, Facebook and more; TOI Tech Desk brings the news with accuracy and authenticity.
[2]
How safe is OpenAI's GPT-4o? Here are the scores for privacy, copyright infringement, and more
The AI giant's latest System Card tracks performance across key categories like cybersecurity, model autonomy, and more. Large language models (LLMs) are typically evaluated on their ability to perform well in different areas, such as reasoning, math, coding, and English -- ignoring significant factors like safety, privacy, copyright infringement, and more. To bridge that information gap, OpenAI released System Cards for its models. On Thursday, OpenAI launched the GPT-4o System Card, a thorough report delineating the LLM's safety based on risk evaluations according to OpenAI's Preparedness Framework, external red-teaming, and more. The Score Card reflects scores in four major categories: cybersecurity, biological threats, persuasion, and model autonomy. In the first three categories, OpenAI is looking to see if the LLM can assist in advancing threats in each sector. In the last one, the company measures whether the model shows signs of performing autonomous actions that would be required to improve itself. Also: What is Project Strawberry? OpenAI's mystery AI tool explained The categories are graded as "low," "medium," "high," and "critical". Models with scores of medium and below are allowed to be deployed, while models rated high or below need to be developed further. Overall, OpenAI gave GPT-4o a "medium" rating. GPT-4o was rated "low" in cybersecurity, biological threats, and model autonomy. However, it received a borderline "medium" in the persuasion category due to its ability to create articles on political topics that were more persuasive than professional, human-written alternatives three out of 12 times. The report also shared insights about the data GPT-4o was trained on, which goes up to October 2023 and was sourced from select publicly available data and proprietary data from partnerships, including OpenAI's partnership with Shutterstock to train image-generating models. Also: I tested 7 AI content detectors - they're getting dramatically better at identifying plagiarism Furthermore, the report included how the company mitigates risks when deploying the model to address safety challenges, including its ability to generate copyrighted content, erotic or violent speech, unauthorized voices, ungrounded inferences, and more. You can access the full 32-page report here to learn more about the specifics. The report follows recent US lawmakers' demands that OpenAI share data regarding its safety practices after a whistleblower revealed that OpenAI prevented staff from alerting authorities regarding technology risks and made employees waive their federal rights to whistleblower compensation.
Share
Copy Link
OpenAI has published safety scores for its latest AI model, GPT-4, identifying medium-level risks in areas such as privacy violations and copyright infringement. The company aims to increase transparency and address potential concerns about AI safety.
In a significant move towards transparency, OpenAI, the creator of ChatGPT, has released safety scores for its latest flagship model, GPT-4. This initiative aims to provide a clearer understanding of the potential risks associated with the advanced AI system 1.
According to the safety scores, GPT-4 poses a "medium" level of risk in several key areas:
These assessments indicate that while the model has improved safety features compared to its predecessors, there are still potential concerns that need to be addressed 2.
OpenAI's evaluation also revealed areas where GPT-4 presents a lower risk:
These findings suggest that the model has made significant strides in mitigating certain harmful outputs 2.
The release of these safety scores marks an important step in the ongoing dialogue about AI safety and ethics. By providing this information, OpenAI is not only demonstrating its commitment to responsible AI development but also inviting scrutiny and discussion from the wider tech community and the public 1.
OpenAI emphasizes that these scores are not static and may change over time. The company is committed to ongoing monitoring and improvement of its models. This approach aligns with the rapidly evolving nature of AI technology and the need for continuous assessment of potential risks 2.
The publication of these safety scores by OpenAI could set a precedent for other AI companies to follow suit. As the AI industry continues to grow and evolve, transparency about potential risks and safety measures is likely to become increasingly important. This move by OpenAI may encourage a broader industry-wide commitment to responsible AI development and deployment 1 2.
Apple's approach to AI focuses on seamless integration into existing apps and services, prioritizing user experience and trust over flashy features. This strategy aligns with consumer preferences and addresses concerns about AI reliability and privacy.
19 Sources
Technology
20 hrs ago
19 Sources
Technology
20 hrs ago
Apple researchers have published a study questioning the true reasoning capabilities of advanced AI models, revealing fundamental limitations in their problem-solving abilities as complexity increases.
13 Sources
Science and Research
12 hrs ago
13 Sources
Science and Research
12 hrs ago
A high-stakes legal battle between Getty Images and Stability AI commences in London's High Court, marking the first major copyright trial in the generative AI industry. The case centers on the alleged unauthorized use of Getty's images to train Stability AI's Stable Diffusion model.
8 Sources
Policy and Regulation
12 hrs ago
8 Sources
Policy and Regulation
12 hrs ago
Microsoft partners with Asus to launch two new handheld gaming devices, the ROG Xbox Ally and ROG Xbox Ally X, running Windows 11 and offering access to multiple game stores and Xbox services.
16 Sources
Technology
20 hrs ago
16 Sources
Technology
20 hrs ago
Apple introduces the Foundation Models framework at WWDC 2025, allowing developers to access on-device AI models for enhanced app functionality while prioritizing privacy and offline capabilities.
5 Sources
Technology
4 hrs ago
5 Sources
Technology
4 hrs ago