OpenAI secures key partnership with Reddit


This allows OpenAI to incorporate conversations from Reddit into ChatGPT and other new products, echoing a previous agreement that the platform had with Google, reportedly valued at $60 million.

Date: 5/28/2024 9:52:14 AM

OpenAI has secured a deal to access real-time content from Reddit through the platform’s data API. 

This allows OpenAI to incorporate conversations from Reddit into ChatGPT and other new products, echoing a previous agreement that the platform had with Google, reportedly valued at $60 million.

The partnership enables OpenAI to better sample the datasets on which their models are trained, allowing AI systems to become more precise and context-aware. For human communication and natural language processing, this means models like ChatGPT can stay continually updated with one of the vastest collections of public discourse available, enabling them to respond more effectively.

As part of this collaboration, Reddit will be able to develop and release new AI-powered tools for its users and moderators, utilising OpenAI’s advanced language models. This collaboration could result in more effective moderation tools and a set of features specifically designed to help users better understand thread information. Features might include summarising content or assisting users in forming responses to replies without having to write everything from scratch.

The primary goal of these features is to refine language interactions for all users. Moreover, as part of this partnership, OpenAI will serve as an advertising partner, enabling Reddit to offer ads that are more tailored and relevant, utilising OpenAI’s capacity to capture the subtleties of user behaviour.

The reaction of Reddit’s community to this partnership remains uncertain, but their historical engagement and vocal opposition to unfavourable executive decisions, such as those during the protests over API pricing, suggest they may respond vigilantly. The acceptance of this partnership will hinge critically on OpenAI’s ability to maintain user privacy and adhere to Reddit’s platform norms.

From OpenAI’s viewpoint, partnering with Reddit signifies a crucial strategic development. It positions the company to highlight its prominent AI technology in direct competition with giants such as Google and Microsoft, and crucially, within the integral realm of social media. For Reddit, this collaboration could provide a substantial advantage over less progressive platforms, potentially reshaping its image and attracting more users.

The partnership offers great promise but also raises critical ethical and methodological questions. Integrating real-time user-generated data to advance AI capabilities could lead to privacy violations and potentially restrict users’ freedom of expression. Additionally, this application of AI might clash with existing ethical norms.

Steve Huffman, the CEO of Reddit, supports the integration, stating it will promote more relevant content and improve community engagement, consistent with the vision of a connected internet. However, navigating the implications of this deal is complex, particularly in light of Reddit’s history with data scraping issues and recent copyright disputes.

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

Tags: ai, artificial intelligence, generative ai, natural language processing, open ai

VIEW COMMENTS
LEAVE A COMMENT
GPT-4o delivers human-like AI interaction with text, audio, and vision integration

About the AuthorBy Ryan Daws | May 14, 2024 https://twitter.com/gadget_ry
Categories: Applications, Artificial Intelligence, Chatbots, Companies, Development, Enterprise, Ethics & Society, Virtual Assistants,
Ryan Daws is a senior editor at TechForge Media with over a decade of experience in crafting compelling narratives and making complex topics accessible. His articles and interviews with industry leaders have earned him recognition as a key influencer by organisations like Onalytica. Under his leadership, publications have been praised by analyst firms such as Forrester for their excellence and performance. Connect with him on X (@gadget_ry) or Mastodon (@gadgetry@techhub.social)
OpenAI has launched its new flagship model, GPT-4o, which seamlessly integrates text, audio, and visual inputs and outputs, promising to enhance the naturalness of machine interactions.

GPT-4o, where the “o” stands for “omni,” is designed to cater to a broader spectrum of input and output modalities. “It accepts as input any combination of text, audio, and image and generates any combination of text, audio, and image outputs,” OpenAI announced.

Users can expect a response time as quick as 232 milliseconds, mirroring human conversational speed, with an impressive average response time of 320 milliseconds.

Pioneering capabilities
The introduction of GPT-4o marks a leap from its predecessors by processing all inputs and outputs through a single neural network. This approach enables the model to retain critical information and context that were previously lost in the separate model pipeline used in earlier versions.

Prior to GPT-4o, ‘Voice Mode’ could handle audio interactions with latencies of 2.8 seconds for GPT-3.5 and 5.4 seconds for GPT-4. The previous setup involved three distinct models: one for transcribing audio to text, another for textual responses, and a third for converting text back to audio. This segmentation led to loss of nuances such as tone, multiple speakers, and background noise.

As an integrated solution, GPT-4o boasts notable improvements in vision and audio understanding. It can perform more complex tasks such as harmonising songs, providing real-time translations, and even generating outputs with expressive elements like laughter and singing. Examples of its broad capabilities include preparing for interviews, translating languages on the fly, and generating customer service responses.


Nathaniel Whittemore, Founder and CEO of Superintelligent, commented: “Product announcements are going to inherently be more divisive than technology announcements because it’s harder to tell if a product is going to be truly different until you actually interact with it. And especially when it comes to a different mode of human-computer interaction, there is even more room for diverse beliefs about how useful it’s going to be.

“That said, the fact that there wasn’t a GPT-4.5 or GPT-5 announced is also distracting people from the technological advancement that this is a natively multimodal model. It’s not a text model with a voice or image addition; it is a multimodal token in, multimodal token out. This opens up a huge array of use cases that are going to take some time to filter into the consciousness.”

Performance and safety
GPT-4o matches GPT-4 Turbo performance levels in English text and coding tasks but outshines significantly in non-English languages, making it a more inclusive and versatile model. It sets a new benchmark in reasoning with a high score of 88.7% on 0-shot COT MMLU (general knowledge questions) and 87.2% on the 5-shot no-CoT MMLU.

The model also excels in audio and translation benchmarks, surpassing previous state-of-the-art models like Whisper-v3. In multilingual and vision evaluations, it demonstrates superior performance, enhancing OpenAI’s multilingual, audio, and vision capabilities.


OpenAI has incorporated robust safety measures into GPT-4o by design, incorporating techniques to filter training data and refining behaviour through post-training safeguards. The model has been assessed through a Preparedness Framework and complies with OpenAI’s voluntary commitments. Evaluations in areas like cybersecurity, persuasion, and model autonomy indicate that GPT-4o does not exceed a ‘Medium’ risk level across any category.

Further safety assessments involved extensive external red teaming with over 70 experts in various domains, including social psychology, bias, fairness, and misinformation. This comprehensive scrutiny aims to mitigate risks introduced by the new modalities of GPT-4o.

Availability and future integration
Starting today, GPT-4o’s text and image capabilities are available in ChatGPT—including a free tier and extended features for Plus users. A new Voice Mode powered by GPT-4o will enter alpha testing within ChatGPT Plus in the coming weeks.


Developers can access GPT-4o through the API for text and vision tasks, benefiting from its doubled speed, halved price, and enhanced rate limits compared to GPT-4 Turbo.

OpenAI plans to expand GPT-4o’s audio and video functionalities to a select group of trusted partners via the API, with broader rollout expected in the near future. This phased release strategy aims to ensure thorough safety and usability testing before making the full range of capabilities publicly available.

“It’s hugely significant that they’ve made this model available for free to everyone, as well as making the API 50% cheaper. That is a massive increase in accessibility,” explained Whittemore.

OpenAI invites community feedback to continuously refine GPT-4o, emphasising the importance of user input in identifying and closing gaps where GPT-4 Turbo might still outperform.