Last Updated on March 4, 2023 by Editorial Team
Author(s): Towards AI Editorial Team
Originally published on Towards AI.
What happened this week in AI by Louis
This week we were pleased to note an acceleration in progress toward open-source alternatives to ChatGPT as well as signs of increased flexibility in access to these models. While many major tech companies are building their own alternative to ChatGPT, we are particularly excited to see open-source alternatives that can make next-generation LLM models more accessible, flexible, and affordable for the machine learning community. We are seeing progress in the release of more open-source foundational LLM models as well as building out of open-source data sets and workflows, together with new innovations to reduce the cost of fine-tuning these models with human feedback.
The strategic partnership between Hugging Face and Amazon Web Services (AWS) looks like a positive step in this direction and should increase the availability of open-source data sets and models hosted on Hugging Face. We were also pleased to see the release of Meta’s LLaMA, 4 foundation models ranging from 7B to 65B parameters. Another example is Laion’s project, which involves crowdsourcing annotations for its OpenAssistant ChatGPT replication project. Carper has also developed open-source RLHF workflows, which range from human annotation with CHEESE to RLHF training with trlX. We are also seeing new approaches that can reduce the human feedback requirements and barriers to entry to developing a ChatGPT-like product — such as Anthropic AI’s constitutional AI approach (behind its Claude model) requiring minimal human labels.
We hope these open-source models and competition can also put pressure on OpenAI to keep costs affordable and to increase the flexibility of interaction with their models. We see signs of this with a glimpse leaked of its new OpenAI Foundry product — a platform for running OpenAI models on a dedicated capacity. This platform will also allow more robust fine-tuning options for its latest models. Also of note in this release is that the future models (GPT-4?) look to offer 32,000 max context length (up from 4,000 today) which will likely bring a lot of new capabilities.
OpenAI is launching a new developer platform that lets customers run the company’s newer machine learning models, like GPT-3.5, on a dedicated capacity. OpenAI describes the forthcoming offering, called Foundry, as “designed for cutting-edge customers running larger workloads.”
Generative AI could potentially become a powerful tool for businesses, providing a new basis for competitive advantage. Enterprises should consider experimenting with generative AI by identifying existing processes that can be enhanced with this technology.
This article features an interview with Professor David Gunkel, discussing the issue of what rights robots, including AI chatbots, should have. The discussion centers on the concept of robot rights, including the background and articulation of rights for AI.
The strategic partnership between Hugging Face and Amazon Web Services (AWS) is expected to make AI open and accessible to everyone. Together, the two companies aim to accelerate the availability of next-generation machine learning models by making them more accessible, efficient, and affordable for the machine learning community.
The use of artificial intelligence for drug discovery has shown promise in the development of potential treatments for opioid addiction. Preclinical studies suggest that blocking kappa-opioid receptors may be an effective approach to treating opioid dependence. AI can be used to optimize and create new drugs that can block the activity of the protein responsible for kappa-opioid receptors, making the drug discovery process more cost-effective and efficient.
Three 5-minute reads/videos to keep you learning
The guide provides a simple explanation of text-to-image models and their use of diffusion to create images from natural language. It also introduces various tools for controlling and improving image generation processes, including ControlNET, ControlNET Pose, and ControlNET LORA.
This post provides a high-level explanation of the inner workings of GitHub’s new code search and offers a glimpse into the system architecture and technical underpinnings of the product. It also discusses how the search functionality allows users to find, read, and navigate code more efficiently.
This is a practical course on Data-Centric AI, focusing on the impactful aspects of real-world ML applications. The class covers algorithms for finding and fixing common issues in ML data, as well as constructing better datasets, with a concentration on data used in supervised learning tasks such as classification.
This guide shares an experience of using AI in education to complete tasks such as generating ideas, producing written material, creating apps, and generating images. The article details the success, changes, and process of using AI in education.
This guide explores the use of AI as a creative tool for essay writing. It discusses how to incorporate AI into writing practices by leveraging it to organize thoughts, capture a voice, summarize complex ideas, assist with idea generation, and evaluate writing quality.
Papers & Repositories
A collection of foundation language models ranging from 7B to 65B parameters by Meta.
This paper proposes a fine-tuning method for aligning text-to-image models using human feedback by collecting human feedback, training a reward function that predicts human feedback, and fine-tuning the model by maximizing the reward-weighted likelihood.
This paper considers an alternative approach to RLHF: converting feedback to instruction by relabeling the original one and training the model for better alignment in a supervised manner.
This work aims to investigate whether strong information extraction (IE) models can be created by directly prompting large language models (LLMs). Specifically, it transforms the zero-shot IE task into a multi-turn question-answering problem using a two-stage framework called ChatIE, which leverages ChatGPT for entity-relation triple extraction, named entity recognition, and event extraction.
The paper offers a thorough evaluation of GPT models for machine translation. It covers various aspects, including the quality of different GPT models compared to state-of-the-art research and commercial systems, the impact of prompting strategies, the resilience to domain shifts, and document-level translation.
Enjoy these papers and news summaries? Get a daily recap in your inbox!
The Learn AI Together Community section!
Upcoming Community Events
The Learn AI Together Discord community hosts weekly AI seminars to help the community learn from industry experts, ask questions, and get a deeper insight into the latest research in AI. Join us for free, interactive video sessions hosted live on Discord weekly by attending our upcoming events.
This week’s session in the (free) nine-part Neural Networks Architectures series led by Pablo Duboue (DrDub) focuses on Multimedia Processing Networks. During this session, he will explore VQA, NMNs, Hierarchical co-attention, Dall-E, Imagen, and Stable Diffusion. Find the link to the seminar here or add it to your calendar here.
Date & Time: 28th February, 11 pm EST
Meme of the week!
Meme shared by friedliver#0614
Featured Community post from the Discord
Carl#1372 has launched a text-to-image AI art generator that enables users to combine different styles or create customized ones to produce unique artwork. It is a free tool designed for novice users to experiment with art. You can check it out here and support a fellow community member. Join the conversation and share your feedback in the thread here.
AI poll of the week!
TAI Curated section
Article of the week
The Microsoft team created an exceptional language model called BioGPT designed to answer medical questions. The researchers claim that BioGPT is as good as a human expert in answering these queries. The article delves into how the Microsoft team trained BioGPT to become such an intelligent language model.
Our must-read articles
If you want to publish with Towards AI, check our guidelines and sign up. We will publish your work to our network if it meets our editorial policies and standards.
Interested in sharing a job opportunity here? Contact [email protected].
If you are preparing your next machine learning interview, don’t hesitate to check out our leading interview preparation website, confetti!
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.
Published via Towards AI