Volume 10, Issue 2 (3-2014)                   JSDP 2014, 10(2): 69-86 | Back to browse issues page

XML Persian Abstract Print

Abstract:   (8688 Views)
Text tokenization is the process of tokenizing text to meaningful tokens such as words, phrases, sentences, etc. Tokenization of syntactical phrases named as chunking is an important preprocessing needed in many applications such as machine translation information retrieval, text to speech, etc. In this paper chunking of Farsi texts is done using statistical and learning methods and the grammatical characteristics of Farsi texts. Many features and labeling methods are examined one by one and the best features and labeling techniques are used for the detection of syntactic phrases and their boundaries. Several machine learning techniques including Support Vector Machine and Conditional Random Fields are used as classifier in our experiments. The impact of the size of training texts on chunking performance was studied as well. Using the proposed methods in this paper, a performance of 84.02% was obtained for detection of phrase boundaries and 78.04% for detection of both phrase boundaries and phrase type
Full-Text [PDF 2581 kb]   (2062 Downloads)    
Type of Study: Research | Subject: Paper
Received: 2013/06/5 | Accepted: 2013/09/10 | Published: 2014/04/8 | ePublished: 2014/04/8