The NLP Cypher | 05.02.21
Last Updated on July 24, 2023 by Editorial Team
Author(s): Ricky Costa
Originally published on Towards AI.
NATURAL LANGUAGE PROCESSING (NLP) WEEKLY NEWSLETTER
The NLP Cypher U+007C 05.02.21
The NLP Index
As an applied machine learning engineer (aka hacker U+1F468βU+1F4BB aka flying ninja U+1F431βU+1F464), Iβm consistently looking for better and faster ways to stay on top of the deep learning and software development circuit. After comparing various sources for research, code, and apps. Iβve discovered that a significant amount of awesome NLP code is not on arXiv and not all NLP research is on GitHub. To obtain a wider scope of current NLP research and code, Iβve created the NLP Index! A search-as-you-type search engine containing over 3,000 NLP repositories (updated weekly) U+1F525. The index contains the research paper, a ConnectedPapers link for a graph of related papers, and its GitHub repo.
The NLP Index
Top NLP Code Repositories – Quantum Stat
index.quantumstat.com
The intent of this platform is for researchers and hackers to obtain information quickly and comprehensively about all things NLP. And not just from research papers, but from awesome apps that are created on top of this research.
Weβve included the option of open search (as opposed to exclusively only serving pre-defined categories) because of inter-dependencies among subject areas. Meaning, sometimes a paper/repo can be both about βknowledge graphsβ and βdatasetsβ simultaneously and itβs difficult to discretize topics. We prefer giving the user the option of openly searching the database across all domains/sectors simultaneously. We also included pre-defined queries with dozens of topics in NLP via the sidebar for convenience.
The index has several attributes such as: search as you type, typo tolerance, and synonym detection.
Synonym Detection
For example, if you search for βdatasetβ the database will also search for βcorpusβ and βcorporaβ text simultaneously to make sure every asset is searched. U+1F91F
Typo Tolerance
If you search βgpt2" it will also include βgpt-2"
Search as you type
It will output results on every character as you type in real-time taking only a couple milliseconds. (thank you memory mapping U+1F648)
Also want to mention that the Big Bad NLP Database has already been merged with the NLP Index! For the most up-to-date compendium of NLP datasets, you can go to the βdataβ section of the sidebar and click dataset or openly search for a specific dataset/task. Eventually, I will sunset the BBND URL and eventually redirect it to the Index.
Want to thank all of the support Iβve received over the past week after taking the NLP Index live. Thank you to Philip Vollet for sharing his dataset with hundred of NLP repos. You can find his posts in the βUnchartedβ section.
More features coming soon. Stay tuned. U+1F649
BERT, Explain Yourself!
Discover why BERT makes an inference using SHAP (SHapley Additive exPlanations); a game theoretic approach to explain the output of any machine learning model. It leverages the Transformers pipeline.
ml6team/quick-tips
It has been over two years since transformer models took the NLP throne U+1F3C5, but up until recently they couldn't tellβ¦
github.com
Colab of the Week
Google Colaboratory
Edit description
colab.research.google.com
Explainable AI Cheat Sheet
Includes graphic, YouTube vid, and several links with papers/ books discussing the topic of explainable AI.
Explainable AI Guide
A brief overview of the Explainable AI cheat sheet with examples.
ex.pegg.io
StyleCLIP is Too Much Fun!
Awesome introduction from Max Woolf on using StyleCLIP (via Colab notebooks) to manipulate headshot pics via text prompts. You can even add your own pictures, the quality is pretty good. For example, take a look at the generation after the text prompt: βFace after using the NLP indexβ U+1F447 U+1F62DU+1F62D
Easily Transform Portraits of People into AI Aberrations Using StyleCLIP U+007C Max Woolf's Blog
GANs, generative adversarial networks, are all the rage nowadays for creating AI-based imagery. You've probably seenβ¦
minimaxir.com
Software Updates
AdapterHub
New version includes BART and GPT-2 models U+1F6A8
Adapters for Generative and Seq2Seq Models in NLP
Adapters are becoming more and more important in machine learning for NLP. For instance, they enable us to efficientlyβ¦
adapterhub.ml
BERTopic
(semi-)supervised topic modeling by leveraging supervised options in UMAP
model.fit(docs, y=target_classes)
Backends:
- Added Spacy, Gensim, USE (TFHub)
- Use a different backend for document embeddings and word embeddings
- Create your own backends with
bertopic.backend.BaseEmbedder
- Click here for an overview of all new backends
Calculate and visualize topics per class
- Calculate:
topics_per_class = topic_model.topics_per_class(docs, topics, classes)
Visualize: topic_model.visualize_topics_per_class(topics_per_class)
Release Major Release v0.7 Β· MaartenGr/BERTopic
The two main features are (semi-)supervised topic modeling and several backends to use instead of Flair andβ¦
github.com
Repo Cypher U+1F468βU+1F4BB
A collection of recently released repos that caught our U+1F441
Gradient-based Adversarial Attacks against Text Transformers
A general-purpose framework, GBDA (Gradient-based Distributional Attack), for gradient-based adversarial attacks, and apply it against transformer models on text data.
facebookresearch/text-adversarial-attack
Install HuggingFace dependences conda install -c huggingface transformers pip install datasets (Optional) For attacksβ¦
github.com
Connected Papers U+1F4C8
Easy and Efficient Transformer
Pytorch inference plugin for transformers with large model sizes and long sequences. Currently supports GPT-2 and BERT models.
NetEase-FuXi/EET
EET(Easy and Efficient Transformer) is an efficient Pytorch inference plugin focus on Transformer-based models withβ¦
github.com
Connected Papers U+1F4C8
MDETR: Modulated Detection for End-to-End Multi-Modal Understanding
Code and links to pre-trained models for MDETR (Modulated DETR) for pre-training on data having aligned text and images with box annotations, as well as fine-tuning on tasks requiring fine grained understanding of image and text.
ashkamath/mdetr
This repository contains code and links to pre-trained models for MDETR (Modulated DETR) for pre-training on dataβ¦
github.com
Connected Papers U+1F4C8
XLM-T β A Multilingual Language Model Toolkit for Twitter
Continues pre-training on a large corpus of Twitter in multiple languages on the XLM-Roberta-Base model. Includes 4 colab notebooks.
cardiffnlp/xlm-t
This is the XLM-T repository, which includes data, code and pre-trained multilingual language models for Twitter. Asβ¦
github.com
Connected Papers U+1F4C8
FRANK: Factuality Evaluation Benchmark
A typology of factual errors for fine-grained analysis of factuality in summarization systems.
artidoro/frank
This repository contains the data for the FRANK Benchmark for factuality evaluation metrics (see our NAACL 2021 paperβ¦
github.com
Connected Papers U+1F4C8
Legal Document Similarity
A collection of state-of-the-art document representation methods for the task of retrieving semantically related US case law. Text-based (e.g., fastText, Transformers), citation-based (e.g., DeepWalk, PoincarΓ©), and
hybrid methods were explored.
malteos/legal-document-similarity
Implementation, trained models and result data for the paper Evaluating Document Representations for Content-basedβ¦
github.com
Connected Papers U+1F4C8
Dataset of the Week: Shellcode_IA32 U+1F469βU+1F4BB
What is it?
Shellcode_IA32 is a dataset containing 20 years of shellcodes from a variety of sources is the largest collection of shellcodes in assembly available to date. This dataset consists of 3,200 examples of instructions in assembly language for IA-32 (the 32-bit version of the x86 Intel Architecture) from publicly available security exploits. Dataset is used for automatically generating shell code (code generation task). Assembly programs used to generate shellcode from exploit-db and from shell-storm were collected.
Where is it?
dessertlab/Shellcode_IA32
Shellcode_IA32 is a dataset consisting of challenging but common assembly instructions, collected from real shellcodesβ¦
github.com
Every Sunday we do a weekly round-up of NLP news and code drops from researchers around the world.
For complete coverage, follow our Twitter: @Quantum_Stat
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming aΒ sponsor.
Published via Towards AI