'Feels More Human', Say Users as Facebook Open-Sources Its Blender Chatbot

  • FAIR claims that Blender, which is available in open source on GitHub, is the largest-ever open-domain chatbot.

  • Blender promises to make interactions with conversational AI systems like Alexa, Siri, and Cortana more natural than before.

  • To achieve Blender’s state-of-the-art performance, researchers at FAIR focused on two engineering steps: blending skills and generation strategy.


Facebook AI Research (FAIR), Facebook’s AI and machine learning division, today detailed work on a comprehensive AI chatbot framework called Blender. FAIR claims that Blender, which is available in open source on GitHub, is the largest-ever open-domain chatbot and outperforms existing approaches to generating dialogue while “feel[ing] more human,” according to human evaluators.


FAIR says Blender is the culmination of years of research to combine empathy, knowledge, and personality into one system. To this end, the underlying models — which benefit from improved decoding and skill blending techniques — contain up to 9.4 billion parameters (configuration variables that define skill on a given problem), or 3.6 times more than previous systems.

Blender promises to make interactions with conversational AI systems like Alexa, Siri, and Cortana more natural than before, whether in enterprise, industrial, or consumer-facing contexts. That’s because they’re able to ask and answer a wide range of questions; display knowledge about specific topics; and express sentiments like empathy, seriousness, or playfulness as circumstances dictate.


 

Blending skills and generation strategies


To achieve Blender’s state-of-the-art performance, researchers at FAIR focused on two engineering steps: blending skills and generation strategy.


“Blending skills” refers to selecting tasks that outperform larger models that lack tuning. As the FAIR researchers point out in a paper, chatbot improvements can be attained by fine-tuning models on data that emphasizes desirable conversational skills. As it turns out, tuning can also minimize undesirable traits learned from large data sets, such as toxicity.


With respect to generation strategy, the choice of decoding algorithm — the algorithm used to generate text from a language model — has an outsized impact on a chatbot’s responses. Because the length of a bot’s responses tend to correspond to human judgments of quality, decoders that strike an appropriate balance are desirable. Responses that are too short are typically perceived as dull or showing a lack of interest, while those that are too long imply waffling or distraction.


Over the course of these engineering steps, the researchers tested three types of model architectures, all of which used Transformers as a base. Transformers — a Google innovation — contain neurons (mathematical functions) arranged in layers that transmit signals from input data and adjust the strength (weights) of each connection, as with all deep neural networks. That’s how they extract features and learn to make predictions, but Transformers also have attention. This means every output element is connected to every input element and the weightings between them are calculated dynamically.


First up was a retriever model that, given a dialogue history (or context) as input, selected the next dialogue response by scoring a large set of candidate responses and outputting the highest-scoring one. The FAIR researchers employed a poly-encoder architecture that encoded features of the context using representations attended to by each candidate response, which they say resulted in improved performance while remaining “tractable” to compute, compared with other architectures, like cross-encoders.


The second model was a generator that produced responses rather than retrieving them from a fixed set. Three models were considered by size, ranging from 90 million parameters to 2.7 billion parameters to 9.4 billion parameters.


The third model attempted to address issues with the generator, namely its tendency to synthesize repetitive responses and to “hallucinate” knowledge. It took a “retrieve and refine” (RetNRef) approach, where the above-described retrieval model produced a response when provided a dialogue history, which was then appended to the input sequence of the generator. In this way, the generator learned when to copy elements of responses from the retriever and when not to so it could output more interesting, engaging, and “vibrant” responses. (Retriever models produce human-written responses that tend to include more vibrant language than standard generative models.)


The FAIR team paired a Wizard Generative model with another retriever that together determined when to incorporate knowledge into chatbot responses. The two models produce a set of initial knowledge candidates and then rank those candidates, after which they select a single sentence and use it to condition response generation. A classifier chooses whether to perform retrieval or not on a per-dialogue basis so as to avoid serving knowledge when it’s not required.


Read More: FACEBOOK LAUNCHES A MESSENGER HUB TO INFORM USERS ABOUT CORONAVIRUS

 

Decoding


For the generative models, the FAIR researchers used a beam search decoder method to generate responses to given dialogue contexts. Beam search maintains a set of partially decoded sequences, called hypotheses, that are appended to form sequences and then scored so the best sequences bubble to the top.


To control the length of the chatbot’s responses, the FAIR team considered two approaches: a hard constraint on the minimum generation length and a classifier that predicted the length of responses and set the minimum generation length constraint to its corresponding prediction. The latter was more complex but resulted in variable-length responses to questions, ensuring the chatbot served long responses when they seemed appropriate.

 

Training the models


To prep the various models that make up Blender, the researchers first performed pretraining, a step that conditions machine learning models for particular tasks. They used Facebook’s own Fairseq, a toolkit that supports the training of custom language models, with data samples from a Reddit corpus containing 1.5 billion comments (with two sets of 360,000 comments each reserved for validation and testing) pruned for known bots, non-English subreddits, deleted comments, comments with a URL, and comments of a certain length.


Next, the FAIR team fine-tuned the models using another Facebook-developed suite — ParlAI — designed for training and testing dialogue models. One training corpus selected was ConvAI2, which contains 140,000 utterances involving paired volunteers getting to know each other by asking and answering friendly questions. Another was Empathetic Dialogues, which consists of 50,000 crowdsourced utterances grounded in an emotional situation. Yet another data set — the Wizard of Wikipedia — comprises 194,000 utterances of 1,250 topics, where each conversation begins with a randomly chosen topic and the goal is to display expert knowledge.


A fourth fine-tuning data set — Blended Skill Talk — aimed to blend the previous three sets (ConvAI2, Empathetic Dialogues, and Wizard of Wikipedia) to combine their respective skills during dialogue. Here, 76,000 utterances were collected with a guided and unguided human speaker, where the guided speaker could select utterances suggested by bots trained on the three individual data sets.

 

Evaluations


Post-training, the researchers evaluated Blender’s performance by comparing it with Google’s latest Meena chatbot, a machine learning model with 2.6 billion parameters. Human volunteers were tasked with answering two questions — “Who would you prefer to talk to for a long conversation?” and “Which speaker sounds more human?” — given 100 publicly released and randomized logs from Meena and the same number of logs generated by Blender. In each case, the volunteers were shown series of dialogues between humans paired with the respective chatbots.


The topics of conversation ranged from cooking, music, movies, and pets to yoga, veganism, instruments, and malls — with the Blender models often going into detail when asked and naming relevant stores, bands, movies, actors, pet species, and pet names. In one example, Blender offered a nuanced answer to a question about how Bach compared with Justin Beiber, while a request that Blender write a song indeed yielded lyrics — although nothing particularly poetic.


When presented with chats showing Meena in action and chats showing Blender in action, 67% of the evaluators said the best-performing Blender-powered chatbot — the one with a generative model containing 9.4 billion parameters pretrained on the Blended Skill Talk corpus — sounded more human. About 75% said they’d rather have a long conversation with the 2.7 billion-parameter fine-tuned model than with Meena. And in an A/B comparison between human-to-human and human-to-Blender conversations, the volunteers expressed a preference for models fine-tuned on Blended Skill Talk 49% of the time, while models trained only on public domain conversations were preferred just 36% of the time.


Problematically, further experiments showed that Blender sometimes produced responses in the style of offensive samples from the training corpora — mostly from Reddit comments. The FAIR researchers say that fine-tuning on the Blended Skill Talk data set mitigated this to an extent but addressing it comprehensively would require using an unsafe word filter and a kind of safety classifier.


We’re excited about the progress we’ve made in improving open-domain chatbots,” wrote Facebook in a blog post. “However, building a truly intelligent dialogue agent that can chat like a human remains one of the largest open challenges in AI today … True progress in the field depends on reproducibility — the opportunity to build upon the best technology possible. We believe that releasing models is essential to enable full, reliable insights into their capabilities.”


The pretrained and fine-tuned Blender models with 90 million parameters, 2.7 billion parameters, and 9.4 billion parameters are available on GitHub, along with a script for interacting with the bot (with safety filtering built in). All code for model evaluation and fine-tuning, including the data sets themselves, is available in ParAI.

Read More: FACEBOOK’S RIDE ENCOURAGES AI AGENTS TO EXPLORE THEIR ENVIRONMENTS

Spotlight

Other News
AI Tech

Qlik Launches AI Council to Responsibly Accelerate Enterprise Adoption of AI

Qlik | January 25, 2024

<p> Qlik, a global leader in data analytics and integration, today announces the establishment of its inaugural AI Council &ndash; an initiative that further embeds leading edge, ethical AI development at the heart of the company&rsquo;s mission and industry proposition. By convening a distinguished set of advisors, Qlik will accelerate the responsible development of its AI-driven product portfolio, benefitting from the expertise of some of the world&rsquo;s most prominent AI experts, to help customers use their data to achieve more significant business outcomes.</p> <p> Qlik&rsquo;s Generative AI Benchmark Report found that 31% of senior executives plan to spend over $10 million on generative AI initiatives in the coming year and 79% have already invested in generative AI tools or projects. Despite this enthusiasm, it also found that they understand the need to surround them with the right data strategies to realize their potential. If the data building blocks of AI are not governed properly as it is democratized across the entire workforce, it could present a serious threat to the efficiency and integrity of business operations. The AI Council has been established to help Qlik&rsquo;s customers navigate these challenges and advise on best practices.</p> <p> Members of the Council will work within Qlik to guide the company&rsquo;s R&amp;D direction, inform its product roadmap and ensure its customers&rsquo; use of Qlik&rsquo;s AI is built with responsibility and ethics front of mind. The Council will also educate Qlik leaders and employees on how to harness the full potential of AI, while providing insights into the priorities of business leaders tasked with demonstrating value from AI investment.</p> <p> The AI Council features some of the most renowned subject matter experts from around the world. More information on these members can be found on our Qlik Staige website:</p> <ul> <li> <strong>Nina Schick &ndash; Author, Advisor and Founder of an advisory firm focused on GenAI</strong><br /> A world-leading authority on generative AI, Nina has long been analyzing emerging technology trends for society. With over two decades of geopolitical experience, she has advised global leaders, including Joe Biden, President of the United States, and was articulating her vision of the &lsquo;AI inflexion point&rsquo; years before ChatGPT made AI a global phenomenon.</li> </ul> <ul> <li> <strong>Dr. Rumman Chowdhury &ndash; Responsible AI leader, engineer, auditor and investor</strong><br /> Rumman is a pioneer in the field of applied algorithmic ethics, creating cutting-edge socio-technical solutions for ethical, explainable and transparent AI. She is currently the CEO and founder of Humane Intelligence, a tech nonprofit that builds a community of practice around algorithmic evaluations. She has also served on multiple boards, including the UK Center for Data Ethics and Innovation, and on UN&rsquo;s Broadband Commission for Sustainable Development, Oxford University&rsquo;s Commission on AI and Governance, and Patterns data science journal. Previously, Rumman was the Director of META (ML Ethics, Transparency, and Accountability) team at Twitter, leading a team of applied researchers and engineers to identify and mitigate algorithmic harms on the platform.</li> </ul> <ul> <li> <strong>Kelly Forbes &ndash; Co-Founder and Executive Director, AI Asia Pacific Institute</strong><br /> Kelly sits at the intersection of policy, research and industry, working with leading organizations and governments to address the risks associated with AI through international cooperation. With extensive experience in the Asia-Pacific region, Kelly has conducted research on AI governance, public-private dialogue and government policy issues.</li> </ul> <ul> <li> <strong>Dr. Michael Bronstein &ndash; DeepMind Professor of Artificial Intelligence, University of Oxford</strong><br /> An award-winning academic, Michael was previously Head of Graph Learning Research at Twitter, a professor at Imperial College London and has held visiting appointments at Stanford, MIT, and Harvard. Michael is also a serial entrepreneur, having founded startups such as Novafora, Invision (acquired by Intel in 2012), Videocites and Fabula AI (acquired by Twitter in 2019).</li> </ul> <p> &quot;The formation of Qlik&#39;s AI Council is a strategic leap, reflecting our deep-seated commitment to not just advancing AI, but doing so with ethical integrity and practical applicability,&quot; said Mike Capone, CEO of Qlik. &quot;Our goal is crystal clear: to enable our customers to harness AI in a way that&#39;s not only transformative, but also responsible. By uniting a cadre of AI luminaries, we are sharpening our focus on delivering AI solutions that are not just cutting-edge, but also seamlessly integrated and governed. This initiative is a pivotal chapter in our journey, marking a bold move towards democratizing AI in a manner that is both accessible and aligned with our core mission of driving substantial, data-driven business outcomes.&quot;</p> <p> Data and analytics leaders from around the world can hear from the AI Council at Qlik Connect, which takes place on June 3-5 in Orlando, Florida. At the pre-eminent event for data analytics, integration, and AI, Council members will share their take on the opportunities and challenges for businesses exploring the value of automation in their data strategy. Additional details and event registration is at www.qlikconnect.com</p> <p> &ldquo;I am excited to join Qlik&rsquo;s AI Council and work with some of the greatest minds in AI to optimize how businesses around the world use data,&rdquo; said Rumman Chowdhury, member of Qlik&rsquo;s AI Council. &ldquo;We&rsquo;ve reached an inflection point where innovations like generative AI are impacting the world as the internet did. This is not the time for complacency. &lsquo;Adopting AI&rsquo; is not as simple as some suggest, but getting left behind is a risky game. By taking responsible steps, organizations can enter an era of unprecedented innovation &ndash; I look forward to being able to contribute to this evolution.&rdquo;</p> <p> &ldquo;In working at JBS USA, I recognize the significance of Qlik&#39;s advancements in AI, embodying a responsible and pragmatic approach to enterprise AI development,&rdquo; said Stephanie Robinson, IT Business Intelligence Manager at JBS. &ldquo;Qlik&#39;s dedication to enhancing AI applications aligns with our focus on employing technology to drive substantial business outcomes. We value Qlik&#39;s commitment to ethical AI practices and are optimistic about the beneficial impact this will have on the industry.&rdquo;</p> <p> <strong>About Qlik</strong><br /> Qlik converts complex data landscapes into actionable insights, driving strategic business outcomes. Serving over 40,000 global customers, our portfolio leverages advanced, enterprise-grade AI/ML and pervasive data quality. We excel in data integration and governance, offering comprehensive solutions that work with diverse data sources. Intuitive analytics from Qlik uncover hidden patterns, empowering teams to address complex challenges and seize new opportunities. Our AI/ML tools, both practical and scalable, lead to better decisions, faster. As strategic partners, our platform-agnostic technology and expertise make our customers more competitive.</p> <p> 2024 QlikTech International AB. All rights reserved. All company and/or product names may be trade names, trademarks and/or registered trademarks of the respective owners with which they are associated.</p> <p> The development, release and timing of any product or functionality described herein remain at the sole discretion of Qlik and should not be relied upon in making a purchasing decision.</p>

Read More

AI Tech

PSPDFKit Leads the AI Revolution in Intelligent Document Processing with Release of XtractFlow

PSPDFKit | January 23, 2024

PSPDFKit, a leading document processing and manipulation platform, announces the release of XtractFlow – a groundbreaking intelligent document processing (IDP) engine powered by generative AI. XtractFlow provides advanced automation for large-scale document classification and data extraction across a broad range of formats, with human-level accuracy. Due to the vast and varied document landscape, traditional IDP platforms are inefficient to run, requiring more resources, time and complex processes. Developers and automation project managers may require several days for setup and deployment of document classification, as well as face complexities with data extraction workflows when working with a broad range of formats in order to achieve high levels of accuracy. This is exacerbated by the limitation of using templates and pre-set patterns to extract data. XtractFlow addresses these challenges by simplifying setup and deployment to a single day and using generative AI from OpenAI and Azure in the first release to intelligently identify the document format, classify the types of documents co-mingled in unstructured storage, and consistently extract data, regardless of its location in the document, with human-level accuracy. AI-Powered Features Supports Hundreds of Formats XtractFlow efficiently extracts data from hundreds of document formats, including PDF, JPEG, Office and CAD files, regardless of document complexity. Automate Document Classification With minimal setup, XtractFlow automatically categorizes documents including contracts, legal filings, lab reports, bank statements and more, in high-volume workflows. Developers can easily customize deployment with either the XtractFlow SDK or API. Extract Data Accurately and Effortlessly XtractFlow effortlessly interprets and retrieves the data users need, avoiding extensive coding and the strict rules for data extraction with a no-code approach, and enables a natural-language experience for the end users. "We've always felt we could go far beyond traditional IDP technology — more accuracy and more intelligence — along with less work spent setting and tuning it for specific workflows," says Miloš Đekić, Vice President of Product Management at PSPDFKit. "Generative AI has enabled us to deliver XtractFlow and bring human-level accuracy to document classification and data extraction in a way that significantly accelerates time to value for our customers." XtractFlow supports PSPDFKit customer data security standards with a strict non-storage policy and aligns with global data retention standards, ensuring integrity and security at every step of your applications and business processes. About PSPDFKit PSPDFKit is helping the world innovate beyond paper with its developer tools, API services, and low-code solutions covering the entire document lifecycle from creation, manipulation, real-time collaboration, signing and markup. The company's products cover all major platforms and support a wide range of programming languages and can be deployed on-premise or in the cloud with ease and at any scale. PSPDFKit has earned its developer first reputation by pioneering products that are easily integrated, completely customizable to fit any deployment and workflow, and trusted by startups, SMBs and some of the largest multinational enterprises.

Read More

AI Tech

AI and Big Data Expo North America announces leading Speaker Lineup

TechEx Events | March 07, 2024

AI and Big Data Expo North America announces new speakers! SANTA CLARA, CALIFORNIA, UNITED STATES, February 26, 2024 /EINPresswire.com/ -- TheAI and Big Expo North America, the leading event for Enterprise AI, Machine Learning, Security, Ethical AI, Deep Learning, Data Ecosystems, and NLP, has announced a fresh cohort of distinguishedspeakersfor its upcoming conference at the Santa Clara Convention Center on June 5-6, 2024. Some of the top industry speakers set to take the stage are: - Sam Hamilton - Head of Data & AI – Visa - Dr Astha Purohit - Director - Product (Tech) Ops – Walmart - Noorddin Taj - Head of Architecture and Design of Intelligent Operations - BP - Temi Odesanya - Director - AI Governance Automation - Thomson Reuters - Katie Sanders - Assistant Vice President – Tech - Union Pacific Railroad - Prasanth Nandanuru – SVP - Wells Fargo - Rodney Brooks - Professor Emeritus - MIT These esteemed speakers bring a wealth of knowledge and expertise to an already impressive lineup, promising attendees a truly enlightening experience. In addition to the speakers, theAI and Big Data Expo North Americawill feature a series of presentations covering a diverse range of topics in AI and Big Data exploring the latest innovations, implementations and strategies across a range of industries. Attendees can expect to gain valuable insights and practical strategies from presentations such as: How Gen AI Positively Augments Workforce Capabilities Trends in Computer Vision: Applications, Datasets, and Models Getting to Production-Ready: Challenges and Best Practices for Deploying AI Ensuring Your AI is Responsible and Ethical Mitigating Bias and Promoting Fairness in AI Systems Security Challenges in the Era of Gen AI and Data Science AI for Good: Social Impact and Ethics Selling Data Democratization to Executives Spreading Data Insights across the Business Barriers to Overcome: People, Processes, and Technology Optimizing the Customer Experience with AI Using AI to Drive Growth in a Regulated Industry Building an MLOps Foundation for AI at Scale The Expo offers a platform for exploration and discovery, showcasing how cutting-edge technologies are reshaping a myriad of industries, including manufacturing, transport, supply chain, government, legal sectors, financial services, energy, utilities, insurance, healthcare, retail, and more. Attendees will have the chance to witness firsthand the transformative power of AI and Big Data across various sectors, gaining insights that are crucial for staying ahead in today's rapidly evolving technological landscape. Anticipating a turnout of over 7000 attendees and featuring 200 speakers across various tracks, AI and Big Data Expo North America offers a unique opportunity for CTO’s, CDO’s, CIO’s , Heads of IOT, AI /ML, IT Directors and tech enthusiasts to stay abreast of the latest trends and innovations in AI, Big Data and related technologies. Organized by TechEx Events, the conference will also feature six co-located events, including the IoT Tech Expo, Intelligent Automation Conference, Cyber Security & Cloud Congress, Digital Transformation Week, and Edge Computing Expo, ensuring a comprehensive exploration of the technological landscape. Attendees can choose from various ticket options, providing access to engaging sessions, the bustling expo floor, premium tracks featuring industry leaders, a VIP networking party, and a sophisticated networking app facilitating connections ahead of the event. Secure your ticket with a 25% discount on tickets, available until March 31st, 2024. Save up to $300 on your ticket and be part of the conversation shaping the future of AI and Big Data technologies. For more information and to secure your place at AI and Big Data Expo North America, please visit https://www.ai-expo.net/northamerica/. About AI and Big Data Expo North America: The AI and Big Data Expo North America is a leading event in the AI and Big Data landscape, serving as a nexus for professionals, industry experts, and enthusiasts to explore and navigate the ever-evolving technological frontier. Through its focus on education, networking, and collaboration, the Expo continues to be a beacon for those eager to stay at the forefront of technological innovation. “AI and Big Data Expo North Americais a part ofTechEx. For more information regardingTechExplease see onlinehere.”

Read More

Software

ON24 Unveils its Next Generation Platform, Unleashing a New Era of AI-powered Intelligent Engagement

ON24 | January 24, 2024

Today, ON24 marks the next chapter of its innovation strategy with the launch of its next generation intelligent engagement platform. With AI-powered intelligence at the core, ON24 will now enable enterprises to continuously engage audiences through hyper-personalized experiences that deliver connected insights and drive cost-efficient revenue growth. “AI will fundamentally reshape sales and marketing and reimagine the customer experience. This means that sales and marketing teams must embrace AI to innovate and adapt, or risk being left behind,” says Sharat Sharan, co-founder, President and CEO. “With the launch of the ON24 Intelligent Engagement Platform, we are excited to leverage our unique foundation of first-party data to give our industry-leading customers an AI advantage, so that they can combine best-in-class experiences, personalization and content, to capture and act on connected data and insights at scale to drive revenue growth. And, we believe AI will continue to be a key differentiation for our own business and technology strategy moving forward, powering a new era of intelligent engagement.” Built on its foundation of first-party data, the ON24 Intelligent Engagement Platform combines best-in-class digital experiences, including the company’s flagship webinar, virtual event and content marketing products, with its new AI-powered Analytics and Content Engine (ACE) to provide an integrated go-to-market solution. The platform’s comprehensive set of capabilities will help sales and marketing teams to: Scale personalized experiences: Build best-in-class, branded experiences that reach their prospects and customers at scale and are hyper-personalized for individuals. Automate continuous engagement: Interact with prospects and customers 24/7 through AI-generated content and always-on nurtures. Deliver connected insights: Track audience engagement data across interactions and analyze engagement across key audience segments. Drive revenue growth: Enable data-driven actions across go-to-market teams to generate pipeline and build lifetime customer relationships. “Today’s launch of the ON24 Intelligent Engagement Platform brings our AI innovation together with more than a decade of market leadership, product development and first-party engagement data,” said Jayesh Sahasi, EVP of Product and CTO at ON24. “We believe this will unlock even greater value for our customers, providing AI-powered intelligence that keeps getting smarter, more personalized and more effective over time." ​​About ON24 ON24 is on a mission to help businesses bring their go-to-market strategy into the AI era and drive cost-effective revenue growth. Through its leading intelligent engagement platform, ON24 enables customers to combine best-in-class experiences with personalization and content, to capture and act on connected insights at scale. ON24 provides industry-leading companies, including 3 of the 5 largest global technology companies, 3 of the 5 top global asset management firms, 3 of the 5 largest global healthcare companies and 3 of the 5 largest global industrial companies, with a valuable source of first-party data to drive sales and marketing innovation, improve efficiency and increase business results. Headquartered in San Francisco, ON24 has offices globally in North America, EMEA and APAC. For more information, visit www.ON24.com.

Read More