1 Introduction

Information extraction, among many other applications, heavily relies on named entity recognition (NER). The majority of prior research on NER relies on extracting general information to identify Person (PER), Location (LOC), Organization (ORG), and Time (TIM) [1, 2]. However, to the best of our knowledge, there have been few studies available on extracting product information, even though it appears that it’s significant and beneficial for information extraction in the area of e-commerce systems.

Due to the nature of these e-commerce systems, products are titled in a short and unstructured way [3]. Therefore, product name extraction or tagging from unstructured texts has a number of challenges. First of all, in unstructured text, there is no formal or well-defined sentence structure. Moreover, there are no punctuation rules in this sentence, and it might not use standard grammar or sentence construction rules. Secondly, as there is no set format for producing a product title, different titles or descriptions for the same item are possible because product titles are generally ad hoc pieces of information. Another issue is that product titles have uncommon words since they do not all contain the same word shape, for example, some of them may contain numbers or have rare words.

There are a number of data sets available for named entity recognition; however, they include structured text related to news, Wikipedia entries, or some other blog entries [4,5,6,7,8]. For noisy data, there are also some data sets collected from Twitter and some other social media resources [7, 9]. However, for the product name extraction, there are only a few e-commerce datasets, which are also mostly in English.

Turkish is a complicated morphological language with extensive features. The term “morphological language” refers to a category of language in which the smallest units of meaning in a language, known as morphemes, are combined to generate words [10]. These morphemes can be put together in a variety of ways to make new words or change the meaning of ones that already exist. In a morphological language, words often have multiple forms, depending on their grammatical function in a sentence. The use of morphemes in Turkish allows for a great deal of flexibility in word formation and can lead to complex word structures and rich systems of inflection, where a single word can convey a variety of meanings through changes in its morphological form. Therefore, many computing tasks cannot be completed by relying solely on word forms.

The problem addressed in this research is the NER task within the context of informal or unstructured Turkish text [10, 11]. Specifically, the challenge lies in identifying and classifying product names in text that lacks traditional grammar rules, sentence structure, and may not conform to standard punctuation and language conventions. This problem arises due to the prevalence of such unstructured text in various online contexts, particularly in e-commerce.

In the literature, there are a few studies that try to apply NER methodologies to structured Turkish text. In addition to this, applying the NER approaches to unstructured Turkish text is very challenging problem [12]. This paper addresses the research gap by focusing on the specific challenges of NER in informal Turkish text, with a primary emphasis on the extraction of product names. Our key research question is as follows:

Research Question: How can the Named Entity Recognition (NER) task be effectively addressed in the context of unstructured Turkish text, particularly for the extraction of product names, and how does the proposed deep learning-based NER model, incorporating Bidirectional Long Short-Term Memory and Conditional Random Field (BiLSTM-CRF) and FastText embeddings, compare with state-of-the-art language models like BERT, as well as traditional baseline NER models?

To answer this question, we propose a novel deep learning model that combines BiLSTM architecture with a CRF layer, further enhanced by FastText embeddings. Additionally, we conduct extensive experiments on a Turkish e-commerce dataset, comparing our model’s performance not only with traditional NER models but also with advanced models like BERT. This research aims to provide insights into effective NER solutions for informal Turkish text and contributes to the broader field of NLP and entity recognition. To the best of our knowledge, this is the first deep learning model for product entity recognition using unstructured Turkish text.

The contributions of our work can be listed as follows:

  • This work tackles a significant research gap in NER by focusing on the specific challenges presented by informal text. While existing NER models excel in structured text, they often struggle to perform effectively in unstructured contexts. This study acknowledges this gap and proposes a solution tailored to informal Turkish text.

  • A new deep learning-based model for unstructured Turkish text is proposed. This model combines a Bidirectional Long Short-Term Memory architecture with a Conditional Random Field layer, bolstered by FastText embeddings. This innovative approach is designed to excel in the absence of traditional sentence structure and grammar.

  • The study conducts a thorough performance evaluation of the proposed model. It not only presents the results of the Bidirectional Long Short-Term Memory and Conditional Random Field model with FastText embeddings but also explores alternative embedding approaches, including Word2Vec and Glove. This comprehensive evaluation demonstrates the effectiveness of the proposed model in addressing the research problem.

  • To provide a comprehensive assessment, the proposed model’s performance is compared with state-of-the-art language model (BERT). This comparison allows for a deeper understanding of the model’s relative strengths and capabilities in handling unstructured Turkish text. In addition to the BERT comparison, the research also compares the proposed model’s performance against baseline NER models. This comparative analysis showcases the advantages and potential of the proposed method over existing techniques.

  • The research employs a Turkish e-commerce dataset collected from the internet, which simulates real-world conditions where informal text is prevalent. By focusing on product name extraction, this work has practical applications in improving search functionality and product recommendation systems in e-commerce platforms.

Section 2 reviews recent studies related to NER. The proposed deep BiLSTM-CRF model for unstructured text is described in Sect. 3. The experimental results of the algorithms are compared and discussed in Sect. 4. Our concluding remarks and possible future work directions are presented in the last section.

2 Related works

In the literature, named entity recognition is applied to various domains. Earlier named entity recognition models have used statistical methods and rule-based models. [13] used decision trees in their studies. [14] proposed a semi-supervised sequential labeling approach using statistical methods. In [15], maximum entropy classifiers are used for NER. Conditional random fields are used for NER in [16]. The study proposed in [17] uses statistical methods for entity extraction and recognition.

Nowadays, with recent and rapid development in artificial intelligence, machine learning models and deep neural networks (DNN) have been widely used in NER and NLP. In [18] authors use a BiLSTM-CRF architecture for sequence labeling problem. Similar to this work, in [19] an end-to-end BiLSTM-CRF architecture is proposed for the same task. In [20] Skip-Gram based embedding approach is utilized with BiLSTM-CRF for English text. Seq2Seq model is used in [21] for NER and compared with BiLSTM model. Contextual string embeddings are used to train LSTM model in [22].

NER in Turkish is a popular and also open research area [23]. Recent studies for Turkish named entity recognition have also utilized neural networks [10]. In [24], authors propose a NER model for Turkish legal texts with a custom-made corpus using BiLSTM and CRF. A neural network based model which employs a semi-supervised learning approach is presented in [25]. The work used a CRF layer on top of the decoder and concatenated word, character, and morphological embeddings as encoder inputs in [26] for Turkish language. The authors presented a LSTM based approach with stacked layers of varied depths while combining word embeddings and writing style embeddings (such as all uppercase letters or sentence case letters) as input representations in [27].

For unstructured or noisy text, there are very few studies in the literature. Çelikkaya et al. [28] presented a NER model for real-world data, which is based on a CRF model. They used morphological and lexical features of noisy Turkish text. The authors adopt a baseline approach using CRF and leverage morphological and lexical features influenced by prior work. They apply this approach to the forum data, speech data, and Twitter datasets, testing different preprocessing scenarios, such as normalization and capitalization.

Another CRF-based model is presented in [29], which employs optional distance-based matching. The authors explore two main approaches for addressing the performance drop in informal texts, namely adapting systems to the characteristics of informal texts or adapting data to suit existing systems. They propose a specialized NER system for tweets without normalization, achieving a 64% F-measure using fundamental features such as word prefixes and suffixes, capitalization, apostrophe information, and gazetteers. The study also emphasizes data preparation techniques, including asciifying datasets and gazetteers with minimal normalization. Furthermore, distance-based matching using the Levenshtein distance algorithm is employed for gazetteer look-up features, and future work aims to enhance these techniques. The paper highlights the importance of tailoring NER systems to the unique properties of informal text domains like tweets, offering valuable insights for NLP applications in this context.

In [30], the authors used morphological and lexical features to utilize the CRF model on Turkish news dataset and also Turkish tweets. The research involves extensive feature engineering to enhance performance on both well-formed texts and user-generated content, introducing new datasets from the Web 2.0 domain and expanding the coverage of named entity types. The approach achieves a promising exact match F1 score of 92% on Turkish news articles and approximately 65% on Web 2.0 datasets. While results are satisfactory for well-formed texts, further research is needed to improve recognition on non-canonical social media content, especially in the case of lowercase proper nouns.

In [31], a rule-based approach has been modeled for Turkish tweets. The study conducts experiments, adapting a rule-based recognition system to better suit Twitter language by relaxing capitalization constraints and expanding lexical resources with diacritics. Additionally, a simplistic tweet normalization scheme is introduced to assess its impact on NER. The findings provide insights into the complexities of NER in Turkish tweets and the effects of tweet normalization, suggesting desirable features for a tailored NER system.

A semi-supervised learning approach based on neural networks is proposed in [32] for Turkish tweets. The study highlights the potential of using in-domain data for unsupervised learning of word embeddings, making it adaptable for morphologically rich languages beyond Turkish.

A BiLSTM-CRF model with different types of embeddings (character, character n-gram, morphological, and orthographic character embeddings) is presented in [11]. The paper explores NER in Turkish noisy text using deep neural networks and transfer learning, as an alternative to rule-based or statistical methods. They utilized another layer to their model which is trained using both noisy and formal text simultaneously. The study tackles the challenges posed by sparse orthography, user style dependencies, and the morphologically rich structure of Turkish. It investigates various word and subword representation techniques without the use of hand-crafted features or external resources. Table 1 summarizes these studies and their F1 score results on noisy data.

Table 1 Related works on noisy Turkish data
Fig. 1
figure 1

Overall structure of proposed model

Fig. 2
figure 2

The structure of BiLSTM-CRF model

3 Proposed work

We introduced an effective, systematic, and practical approach based on a Bidirectional Long Short-Term Memory and Conditional Random Field deep learning model for product entity detection in unstructured text. This method encompasses data scraping, preprocessing, data annotation, feature engineering, deep learning model construction, training, and evaluation phases. Figure 1 provides an overview of our proposed method, outlining the steps involved in detecting entities in unstructured text.

The deep learning model comprises three main layers as shown in Fig. 2: the Embedding layer, the BiLSTM layer, and the CRF layer. Within the Embedding layer, unstructured text undergoes tokenization, and each token is transformed into numerical form using an embedding technique. This work employs two distinct embedding techniques: FastText and BERT-based embeddings. The vector representations of unstructured text are subsequently input into the acrshortbilstm model. The output of the acrshortbilstm layer is then utilized as input for the CRF layer, which serves the purpose of classifying named entities. To facilitate comparison, two distinct BiLSTM-CRF models were developed for each of the embedding techniques.

Fig. 3
figure 3

Sample product page taken from e-commerce website [36]

3.1 Data collection and annotation

The dataset is collected from Turkish e-commerce web site [36]. The product titles are first extracted and cleaned using preprocessing tools. After cleaning the titles, manual labeling performed on each row of the dataset which contains 1167 rows. The snapshot of e-commerce site is shown in Fig. 3. From this product page, the text “Mas Makas Force 220 Mm Paslanmaz Çelik 1222 MAS” is extracted and then the product name in this text is labeled manually which is Makas (Scissors. There is only one type of entity, namely PRODUCT(PRO), in this work and the labels used to annotate the product entities are in IOB2 format. These format contains Inside, Outside and Begin tags and shown in Table 2.

Table 2 IOB2 Format Tags for NER

As an example, the product title “Mas Makas Force 220 Mm Paslanmaz Çelik 1222 MAS” first cleaned and numeric data is removed. After this process, we obtain the clean product title as “Mas Makas Force Mm Paslanmaz Çelik MAS”. For this text we have the tag representations shown in Table 3.

Table 3 Annotations of sample text using IOB2 Format

3.2 Embedding layer

In natural language processing (NLP), the word embedding approach is used to represent words as numerical vectors. This is accomplished by giving each word in a corpus (collection of text) a vector with a fixed size. These vectors are usually generated using a neural network-based method, where the network is trained on huge amount of text data and the word vectors are learned by optimizing a specific objective function. The generated vectors store various aspects of each word’s meaning, including its context, grammar, and semantics.

The ability to manipulate words mathematically by performing operations like vector addition and subtraction on them makes it feasible to gain insightful knowledge about the relationships between words. In several NLP tasks, such as sentiment analysis, named entity recognition, and machine translation, word embeddings have shown to be tremendously helpful. Additionally, they are utilized in a number of well-known deep learning models for NLP, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs). In short, word embedding is simply defined as mapping words to high-dimensional feature vectors.

In the literature, there are popular studies which uses different approaches to represent words in numeric format such as FastText, Glove, and Word2Vec.

FastText is a recent word embedding library developed by Facebook. The n-gram feature of the phrase is used as an additional feature to input by FastText model in addition to the word representation for each word in the sentence. This feature makes FastText as a ideal technique for generating words’ vector representations of morphologically complex language such as Turkish. In this work, we used FastText embeddings as morphological embedding layer and compared it with other two state-of-the-art embedding techniques, Glove and Word2Vec.

In addition to FastText, we also used BERT model as word embedding approach. BERT (Bidirectional Encoder Representations from Transformers), released in late 2018. Traditional word embeddings generate a fixed-length vector representation for each word in a vocabulary, based on its co-occurrence statistics with other words in a large corpus of text. In contrast, BERT uses masked language modeling (MLM) to train a deep bidirectional transformer network on a huge corpus of text to provide embeddings for complete sentences or text sequences. BERT randomly masks some of the input tokens during training, and the model has been assigned with predicting the masked tokens based on the context.

As a result, embeddings that accurately capture the context-dependent meaning of each word in the sentence are created, enabling the model to acquire a deep representation of the relationships between words in a sentence. By layering a task-specific layer on top of the pretrained BERT model and optimizing the entire network on a task-specific dataset, the embeddings can be fine-tuned on a downstream NLP task, such as sentiment analysis or named entity recognition, once the model was successfully trained.

In this work, we use BERT to extract features, namely word embedding vectors, from unstructured text data. BERT can express tokenized words as corresponding word embeddings. BERT is better at handling unlabeled data and the word vector expressed by BERT includes context information in addition to its own information. We utilized a pretrained language mode, namely Turkish BERT (BERTurk) for this purpose.

3.3 Deep learning models

3.3.1 BILSTM

Recurrent neural network (RNN) that can handle the vanishing gradient problem is known as Long Short-Term Memory (LSTM) network. LSTM is also better at maintaining long-range connections and understanding the connection between values at the start and end of a sequence. By modifying a gating structure of a traditional RNN model, the LSTM model can learn or retain a longer data sequence. Therefore, LSTM has three gates: input, forget, and hidden. An LSTM unit with these three gates and a memory cell forms a layer of neural network neurons, and each neuron has a hidden layer and a current state. Figure 4 shows the LSTM cell’s structural layout.

Fig. 4
figure 4

LSTM cell structure

The forget gate is used to specify whether or not certain data will be kept. This preservation is accomplished using the following formula;

$$\begin{aligned} f_t = \sigma (W_f[h_{t-1}, x_t] + b_f) \end{aligned}$$

where \(x_t\) denotes input at time t, \(h_{t-1}\) denotes the output of previous cell, and \({\sigma }\) is a sigmoid function. If a forget gate outputs 1 (one), the information is stored in the cell state. The sigmoid function creates a vector in the following step. New possible values are stored in this vector. The updated values are specified by input gates, and the vector \(C^{'}_t\) is updated with possible new values. This new vector is evaluated with the following formulas;

$$\begin{aligned} i_t= & {} \sigma (W_i[h_{t-1}, x_t] + b_i) \\ C^{'}_t= & {} {\text {tanh}}(W_c[h_{t-1}, x_t] + b_c) \end{aligned}$$

Now cell’s old state \(C_{t-1}\) is updated to new cell state \(C_t\).

$$\begin{aligned} C_t = f_t * C_{t-1} + i_t * C^{'}_t \end{aligned}$$

Eventually, we select the network’s output regarding on the cell state. This selection process is carried out by using the following formulas;

$$\begin{aligned} o_t= & {} \sigma (W_o[h_{t-1}, x_t] + b_o) \\ h_t= & {} o_t * {\text {tanh}}(C_t) \end{aligned}$$

A special type of LSTM network which is used generally for natural language processing is called Bidirectional Long Short-Term Memory. In BiLSTM, there are two different LSTM networks in order to represent the input in both direction (backward and forward). Therefore, representing the input in both directions of the sequence, it is an effective tool for modeling the sequential relationships between words and phrases. In BiLSTM, two hidden states are introduced, one for accessing the context of previous input \(\overleftarrow{h_t}\), and another one for accessing the context of the next input \(\overrightarrow{h_t}\). Therefore, the formula for the hidden state of BiLSTM can be defined as:

$$\begin{aligned} h_t = \overrightarrow{h_t} \oplus \overleftarrow{h_t} \end{aligned}$$

In summary, BiLSTM reverses the direction of information flow by adding one extra LSTM layer. It simply means that in the additional LSTM layer, the input sequence flows backward. The outputs from the two LSTM layers are then combined in a variety of ways, including average, sum, multiplication, and concatenation. BiLSTM process the input sequence of \(S = {w_1, w_2,\ldots ,w_n}\) sequentially. Here \(w_i\) denotes the ith word of the sentence S. The embedding of each word w is supplied to BiLSTM model in order to produce the concatenated output X.

3.3.2 CRF

A Conditional Random Field is a type of Markov random fields and they can be used to predict a specific label or tag using its neighbor labels with some consistency. CRF is generally utilized to predicate these adjacent labels in sequential data. In this work, CRF layer is utilized as top hidden layer of BiLSTM network to concatenate the last hidden states from the underlying network. The CRF layer tries to predict the sequence of labels \(Y = {y_0, y_1,\ldots , y_n}\) of a given sentence \(X = {x_0, x_1,\ldots , x_n}\)by using following equation:

$$\begin{aligned} p(y \mid x) = e^{{\text {Score}}(x,y)}/\sum _{y^{'}} {e^{{\text {Score}}(x,y^{'})}} \end{aligned}$$

In this equation, \(p(y \mid x)\) denotes the conditional probability, and Score is computed using the following equation:

$$\begin{aligned} Score(x,y) = \sum _{i=0}^T{A_{y_{i}, y_{i+1}}} + \sum _{i=1}^T{P_{i,y_{i}}} \end{aligned}$$

where \(A_{y_{i}, y_{i+1}}\) represents probability of transition from label i to label j, and \(P_{i,j}\) denotes the score of the jth label of the word ith. This \(P_{i,j}\) is represented as matrix and it is the output of BiLSTM network. With CRF model, the main goal is to maximize the log of conditional probability \(log(p(y \mid x))\).

3.3.3 BiLSTM-CRF

This architecture has three layers: an embedding layer, a Bidirectional LSTM layer, and a dropout layer, just like the core BiLSTM model. However, it also has the following two extra parts:

  • TimeDistributed(Dense) Layer: This layer applies a dense layer to each time step of the BiLSTM output sequence, allowing the network to learn features at each time step.

  • CRF Layer: This layer applies a CRF algorithm to the output of the TimeDistributed(Dense) layer, which models the dependencies between adjacent output labels and helps to ensure that the predicted label sequence is globally optimal.

Overall, the BiLSTM-CRF model has been proven to be superior to many other state-of-the-art models in sequence labeling tasks such named entity recognition and part-of-speech tagging.

3.4 Evaluation

We employ Precision, Recall, and F1 score as key metrics to assess the effectiveness of our proposed model. Precision gauges the model’s capability to accurately identify relevant entities, while Recall evaluates its ability to capture all relevant entities within a dataset. The F-score, on the other hand, serves as the harmonic mean of Precision and Recall, offering a balanced measure of the model’s overall performance. Detailed calculations for these metrics are provided below.

$$\begin{aligned} {\text {Precision}}= & {} \frac{{\text {TP}}}{{\text {TP}}+{\text {FP}}}\\ {\text {Recall}}= & {} \frac{{\text {TP}}}{{\text {TP}}+{\text {FN}}}\\ F1= & {} \frac{2*{\text {Precision}}*{\text {Recall}}}{{\text {Precision}}+{\text {Recall}}} = \frac{2*{\text {TP}}}{2*{\text {TP}}+{\text {FP}}+{\text {FN}}} \end{aligned}$$

where True Positive (TP) represents entities correctly recognized by the model and that align with annotated entities. False Positive (FP) pertains to entities erroneously identified by the model, which do not align with annotated entities. Conversely, False Negative (FN) encompasses annotated entities that the model fails to recognize.

4 Experiments

In the first set of experiments, we compared performance results of different BiLSTM-CRF models for each embedding approach. We implemented our architecture in Python with TensorFlow and Keras. The dataset contains 1167 labeled rows and split into train and test sets with %80-%20 ratio. All experiments were conducted on a single personal computer equipped with a single NVIDIA RTX 3080 GPU boasting 32 GB of memory.

In the course of our research, we implemented a systematic strategy aimed at fine-tuning the hyperparameters of our deep learning model, a crucial element in attaining exceptional performance levels. To facilitate this process, we employed a methodology referred to as grid search. This approach provided us with the means to comprehensively explore a broad spectrum of hyperparameter combinations, systematically adjusting key parameters such as learning rates, batch sizes, and model architecture configurations. This diligent exploration led to the identification of highly promising hyperparameter configurations that notably bolstered the model’s effectiveness.

The adoption of grid search underscored our unwavering commitment to meticulous parameter optimization, driving us toward the attainment of the most favorable model settings and, ultimately, yielding exceptional outcomes in our study. Hence, the hyperparameter values for BiLSTM model implementation were optimized for all set of experiments as follows: batch size 128; epoch 50; learning rate 0.002; LSTM hidden size 50; and LSTM dropout rate 0.1.

Fig. 5
figure 5

BiLSTM-CRF model with FastText embedding

Table 4 Performance results of proposed work with FastText embedding and different dimensions

First of all, we conducted test on BiLSTM-CRF with FastText embedding layer (Fig. 5). In the embedding layer we used two different dimension number, respectively, 50 and 300. The results in Table 4 show that there is a little performance gain using 300 dimensional embedding vector.

Table 5 Comparison of BiLSTM-CRF model with different embedding layers

In order to show the performance of the proposed work, we utilized BiLSTM-CRF with two other embeddings: Glove and Word2Vec. The performance results are shown in Table 5. When compared to alternative word embedding strategies, using the FastText word embedding model significantly improves model performance.

Fig. 6
figure 6

BiLSTM-CRF model with BERT embedding

Table 6 Performance results of proposed models with FastText and BERT embedding
Fig. 7
figure 7

F1 results of each embedding technique

Fig. 8
figure 8

Accuracy results of each embedding technique

Moreover, we designed second set of experiments to compare the proposed approach with BERT embedding (Fig. 6) and the results are shown in Table 6. This experiment results reveal that using BERT for embedding layer shows very close performance to proposed model. The F1 scores for each model are shown in Fig. 7 and accuracy results for each model are shown in Fig. 8. The experiments reveal that using morphological embedding for unstructured Turkish text outperforms word based embedding.

Fig. 9
figure 9

BERT-CRF model

Fig. 10
figure 10

BiLSTM (without CRF layer) model with FastText embedding

In addition to these set of experiments, we conducted another set of experiments to compare our model performance with those of two baseline methods. In the first baseline method BERT transformer model is utilized with CRF layer on top of this model (BERT-CRF) (Fig. 9). In the second baseline model, a standard BiLSTM model is utilized without CRF layer and FastText embedding is used as embedding layer again (BiLSTM FastText) (Fig. 10).

Table 7 Comparison results of proposed model and baseline models
Fig. 11
figure 11

F1 results of proposed model and baseline models

Fig. 12
figure 12

Accuracy results of proposed model and baseline models

The test results are summarized in Table 7 and overall F1 score and accuracy values for these model and also BiLSTM-CRF with BERT embedding is depicted in Figs. 11 and 12 for overall comparison. The results also confirm that our proposed model has better results than these baseline models. Moreover, using FastText embeddings has better results compared to BERT embeddings. FastText embeddings may be a better choice than BERT embeddings because they are generated based on character-level n-grams, which allows them to capture subword information and handle variations in spelling and morphology. This capability makes them more robust to unstructured text data than traditional word embeddings, which can be easily affected by misspellings, typos, and other errors. In contrast, BERT embeddings are contextualized and rely heavily on the overall structure and meaning of the sentence. Therefore, BERT embeddings are less robust to unstructured text data, as errors or variations in the text can disrupt the overall context and meaning of the sentence.

4.1 Error analysis

To examine the underlying factors contributing to misclassification by the techniques in our corpus, we compiled and synthesized the results of the experiments. To achieve this objective, we conducted a qualitative error analysis to investigate common errors in the results. Typically, product names, which are frequent, are accurately tagged. However, when they are infrequent or written with symbols and numbers, there is a possibility that they may not be tagged correctly. Another typical error occurs when product names consist of multiple words, especially if irrelevant and meaningless words are located between them.

Another error occurs when product names are shortened. For example for the product definition Petlas 205/55 R16 91 H MultiAction PT565 Oto 4 Mevsim Lastiği ( Üretim Yılı: 2022 ) (winter tire product definition and “Oto” is shortened for “Otomobil” which means Automobil or car) “Oto” word is used to refer “Otomobil”. Therefore, the expected tagging is “Otomobil 4 mevsim lastiği” (car four season tire), but the model tagged it as “Mevsim Lastiği”.

These are the common errors, however, the model correctly tags frequent and single-word production names.

5 Conclusion

Dealing with unstructured data is a challenging task in Natural Language Processing (NLP), and there have been several contributions to improving Named Entity Recognition (NER) on this type of data.

In this study, we propose a BiLSTM-CRF model designed for unstructured Turkish text. Our methodology begins by collecting data from Turkish e-commerce websites. Subsequently, the data undergoes a meticulous cleaning process involving the removal of unwanted characters, symbols, noise (such as numbers, punctuation, and special characters), and redundant or irrelevant information.

The transformed input data is then converted into numerical representations utilizing embedding techniques. When handling unstructured textual data, it proves beneficial to employ word embeddings that exhibit resilience to text variations and errors. Our model employs diverse embedding approaches to address these variations and errors. Particularly, we utilize FastText, a widely used library developed by Facebook, to generate word embeddings based on character-level n-grams.

Given that Turkish is a language known for its morphological complexity, the utilization of FastText embeddings allows us to capture subword information. This feature equips the model to effectively manage variations in spelling and morphology. This stands in contrast to conventional word embeddings, which are generated at the word level and may falter in accommodating spelling variations or morphological distinctions among related words.

As an alternative embedding approach, we incorporate BERT embeddings to represent unstructured text. BERT embeddings have demonstrated exceptional performance across a broad spectrum of NLP tasks, even when dealing with noisy or imperfect text data. Their efficacy stems from their ability to grasp intricate contextual relationships between words within a sentence.

We proceed to compare these two embedding techniques with each other and with established traditional embedding methods, all within the framework of a BiLSTM-CRF model. The core of our model lies in the bidirectional Long Short-Term Memory architecture, which adeptly captures the context of the input sequence. Additionally, we incorporate a Conditional Random Field layer to model interdependencies among the output labels. This sophisticated architecture enables the model to predict the label of each token within the input sequence by leveraging both forward and backward contextual information.

We conducted our experiments using real-world datasets, and the results reveal compelling findings. Notably, when applied to Turkish, a language known for its morphological complexity, FastText embeddings outperform traditional embedding techniques. Furthermore, our results indicate that FastText embeddings exhibit a slight performance advantage over BERT embeddings.

In the pursuit of advancing the capabilities of product name recognition in unstructured text, several promising avenues for future research emerge. First and foremost, there is the compelling prospect of extending the model’s proficiency to multilingual contexts. Given the diverse linguistic landscape of e-commerce platforms, exploring the adaptation of our model to different languages could enhance its applicability.

Another promising avenue involves the fine-tuning of the model for specific domains within the e-commerce sector, such as electronics, fashion, or food. Tailoring the model to recognize domain-specific product names can lead to heightened accuracy and relevance in recommendations.

To further refine our approach, advanced data preprocessing techniques warrant exploration. This includes strategies for handling misspelled words and text that departs from standard grammar and syntax, ensuring the model’s robustness in the face of diverse textual idiosyncrasies. Moreover, the implementation of active learning strategies holds potential for reducing the manual annotation burden associated with data collection. By selecting the most informative instances for manual labeling, we can not only enhance model performance but also alleviate annotation costs.

These future research directions collectively pave the way for further innovation and advancement in the realm of product name recognition, with broad-reaching implications for e-commerce and related fields.