Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Read by thought-leaders and decision-makers around the world. Phone Number: +1-650-246-9381 Email: [email protected]
228 Park Avenue South New York, NY 10003 United States
Website: Publisher: https://towardsai.net/#publisher Diversity Policy: https://towardsai.net/about Ethics Policy: https://towardsai.net/about Masthead: https://towardsai.net/about
Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Founders: Roberto Iriondo, , Job Title: Co-founder and Advisor Works for: Towards AI, Inc. Follow Roberto: X, LinkedIn, GitHub, Google Scholar, Towards AI Profile, Medium, ML@CMU, FreeCodeCamp, Crunchbase, Bloomberg, Roberto Iriondo, Generative AI Lab, Generative AI Lab Denis Piffaretti, Job Title: Co-founder Works for: Towards AI, Inc. Louie Peters, Job Title: Co-founder Works for: Towards AI, Inc. Louis-François Bouchard, Job Title: Co-founder Works for: Towards AI, Inc. Cover:
Towards AI Cover
Logo:
Towards AI Logo
Areas Served: Worldwide Alternate Name: Towards AI, Inc. Alternate Name: Towards AI Co. Alternate Name: towards ai Alternate Name: towardsai Alternate Name: towards.ai Alternate Name: tai Alternate Name: toward ai Alternate Name: toward.ai Alternate Name: Towards AI, Inc. Alternate Name: towardsai.net Alternate Name: pub.towardsai.net
5 stars – based on 497 reviews

Frequently Used, Contextual References

TODO: Remember to copy unique IDs whenever it needs used. i.e., URL: 304b2e42315e

Resources

Unlock the full potential of AI with Building LLMs for Production—our 470+ page guide to mastering LLMs with practical projects and expert insights!

Publication

Learn AI Together — Towards AI Community Newsletter #13
Artificial Intelligence   Latest   Machine Learning

Learn AI Together — Towards AI Community Newsletter #13

Author(s): Towards AI Editorial Team

Originally published on Towards AI.

Good morning, fellow AI enthusiasts! In this week’s edition of the Learn AI Together newsletter, we have a comprehensive guide designed to teach everything about large language models (LLMs) in 2024 for free. ‘From zero to hero with LLMs’ is a curated collection of resources to make cutting-edge AI knowledge accessible to all with the wealth of free online materials. We hope you find it useful.

We also wanted to learn more about how many of you are working in the industry, a handy poll for those looking to find the best deal possible. More details in this iteration!

I wish you all a great read and an amazing weekend!

What’s AI Weekly

Louis-François Bouchard has compiled LLM resources as a complete guide to starting and improving your LLM skills in 2024 without an advanced background in the field. It is intended for anyone with a small programming and machine learning background. There is no specific order, but a classic path would be from top to bottom. All resources listed here are free, except some online courses and books, which are recommended for a better understanding. Still, it is possible to become an expert without them, with more time spent on online readings, videos, and practice. Find the ‘From zero to hero with LLMs’ guide here!

  • Louis-François Bouchard, Towards AI Co-founder & Head of Community

Learn AI Together Community section!

Featured Community post from the Discord

Ryios shared an exciting new idea with the community: instead of 1 mega model, one good general language model feeds down into hundreds, thousands, or even millions of smaller micro models that specialize in something. Mega Models are too expensive to train, host, and run inference. They exceed the capabilities of most consumer hardware. To address this, instead of training a model to know everything, ryios proposes training a model to be well-versed in the language. This AI’s task is to translate user prompts into formats that can be understood by other AIs downstream, much like a translator or secretary. Join the conversation and share your thoughts in the thread!

AI poll of the week!

While Google Cloud seems the preferred choice, the thread is flooding with recommendations for Digital Ocean, Oracle, AWS, and more. Let us know your favorites in the thread!

Collaboration Opportunities

The Learn AI Together Discord community is flooding with collaboration opportunities. If you are excited to dive into applied AI, want a study partner, or even want to find a partner for your passion project, join the collaboration channel! Keep an eye on this section, too — we share cool opportunities every week!

1. Mh_aghajany is looking for fellow learners to explore Machine Learning, Deep Learning, and LLM. If you’re passionate about ML and interested in collaborative learning, connect in the thread!

2. Our friends at Zoī are hiring their Chief AI Officer. Zoī is at the crossroads of 3 domains: Medical, Data Science, and BeSci. Zoī aims to create personalized user manuals for each member by gathering only the necessary data to provide tailored recommendations based on thousands of factors. Find more information in the thread!

3. Usmanyousaaf is looking for a study partner to dive into ANN, CNN, RNN, LSTM, GRU, Transformers, pre-trained models, GANs, and more. If you are also learning the math behind each and want to work on projects, reach out in the thread!

Meme of the week!

Meme shared by ghost_in_the_machine

TAI Curated section

Article of the week

Advanced RAG 04: Re-ranking by Florian June

This article introduces RAG’s re-ranking technique and demonstrates how to incorporate re-ranking functionality using two methods. Re-ranking is crucial in the Retrieval Augmented Generation (RAG) process. In a naive RAG approach, a large number of contexts may be retrieved, but not all are necessarily relevant to the question. Re-ranking allows for the reordering and filtering of documents, placing the relevant ones at the forefront, thereby enhancing the effectiveness of RAG.

Our must-read articles

1. easy-explain: Explainable AI with GradCam by Stavros Theocharis

GradCam is a widely used Explainable AI method that has been extensively discussed in forums and literature. Therefore, the author has included this common method in his package, “easy-explain,” but in an abstract way so that anyone can use it easily. The article walks you through it.

2. LangChain 101: Part 3b. Talking to Documents: Embeddings and Vectorstores by Ivan Reznikov

In this part of the LangChain 101 series, the author discusses what embeddings are and how to choose one, what vector stores are, how vector databases differ from other databases, and, most importantly, how to choose one! All code is provided and duplicated in Github and Google Colab.

3. Lumiere, Google’s Amazing Video Breakthrough by Ignacio de Gregorio

Google has taken us one step closer, as their approach to AI video synthesis is not only revolutionary but also showcases incredible video quality and a wide range of amazing skills like video in/outpainting, image animation, and video styling, making it the new reference in the field.

4. Understanding the Mechanics of Neural Machine Translation by Saif Ali Kheraj

As large language models become more prevalent, we must study and concentrate on attention models, which play an essential role in both Transformer and language models. First, it is a good idea to understand the Sequence to Sequence Encoder Decoder Network. After that, proceeding to the most important “Attention Model” and examine it in greater detail.

If you are interested in publishing with Towards AI, check our guidelines and sign up. We will publish your work to our network if it meets our editorial policies and standards.

Think a friend would enjoy this too? Share the newsletter and let them join the conversation.

Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.

Published via Towards AI

Feedback ↓