Please login first
Transformers architecture application in high-quality business names generation
* 1, 2 , 2 , 2 , 2
1  Kaunas University of Technology
2  Zyro Inc
Academic Editor: María N. Moreno García (registering DOI)

The continuous improvement of artificial intelligence/machine learning is leading to an increasing search for the wider application of these technological solutions not only to structured data but also to unstructured ones. In order to apply data science to language processing, an area has emerged - natural language processing (NLP). Natural language processing is the computer analysis and processing of natural language (which can be spoken and written) using a variety of technologies aimed at adapting human language to various tasks or computer programs using linguistic methods.

At present, natural language processing is finding more and more different ways to adapt to real practical problems. These tasks can range from searching for meaningful information in unstructured data (Pande and Merchant, 2018), analyzing sentiments (Yang et al., 2020; Dang et al., 2020; Mishev et al., 2020), and translating the text into another language ( Xia et al., 2019; Gheini et al., 2021) to fully automated human-level text creation (Wolf et al., 2019; Topal et al., 2021). The data set for this study consists of 350,928 observations/business names (299,964 observations in the training sample and 50,964 observations in the test sample). These data were collected using the websites of start-ups from around the world. The aim of this study is to apply natural language modeling models of transformer architecture to generate high-quality business names.

Keywords: Natural language processing; NLP; machine learning; Natural Language Generation; NLG; artificial intelligence
Comments on this paper
Mona Almansoori
Choosing the GPT-2 model used by most researchers for its experimental and surprising capabilities, high performance and excellent quality is a good choice, and as you mentioned it is the best modified model currently.
I wish you success your assessment and results.
Mantas Lukauskas
I would like to thank you for your comment. As we mentioned in our research, we also tested the newest GPT-3 models. Still, these models did not outperform current freely available GPT-2 models like EleutherAI GPT Neo-125M and GPT Neo-1.3B models. Because these models are applicable in the real-world application which needs a lot of speed due to the high demand for smaller models GPT Neo-125M is one of the best solutions there.