2—classification: Transformer. A Transformer is a neural network architecture that learns context from texts. One example is the Bidirectional Encoder Representations from Transformers (BERT language model) used in natural language processing. BERT was introduced in October 2018 by researchers at Google and is implanted by huggingface (Hugging Face, Inc., New York). A transformer is a deep learning architecture based on the multi-head machine learning attention mechanism. Machine learning-based attention is a mechanism which intuitively mimics cognitive attention. It calculates “soft” weights for each word, more precisely for its embedding, in the context window, and is used to train large language model (LLM) datasets. The use of BERT is given as a non-limiting example of transformer architecture, and any suitable transformer architecture can be used.
This is accomplished by email text processing, and content and link detection. Email text processing addresses text patterns and words, and includes text manipulation by converting text to lowercase and stripping out special characters, numbers, and stop words that could obfuscate the message intended to be conveyed by the text. By way of non-limiting examples, this would use a Masked Language Model (MLM) RegEx, and a text detection library such as Pandas, scikit-learn, Re and/or a Natural Language Toolkit (NLTK). Pandas is a software library written for the Python programming language for data manipulation and analysis. In particular, it offers data structures and operations for manipulating numerical tables and time series. Scikit-Learn is a free software machine-learning library for the Python programming language.