<?xml version="1.0" encoding="utf-8"?>
<journal>
<title>Signal and Data Processing</title>
<title_fa>پردازش علائم و داده‌ها</title_fa>
<short_title>JSDP</short_title>
<subject>Engineering &amp; Technology</subject>
<web_url>http://jsdp.rcisp.ac.ir</web_url>
<journal_hbi_system_id>1</journal_hbi_system_id>
<journal_hbi_system_user>admin</journal_hbi_system_user>
<journal_id_issn>2538-4201</journal_id_issn>
<journal_id_issn_online>2538-421X</journal_id_issn_online>
<journal_id_pii></journal_id_pii>
<journal_id_doi>10.61882/jsdp</journal_id_doi>
<journal_id_iranmedex></journal_id_iranmedex>
<journal_id_magiran></journal_id_magiran>
<journal_id_sid>1</journal_id_sid>
<journal_id_nlai>8888</journal_id_nlai>
<journal_id_science></journal_id_science>
<language>fa</language>
<pubdate>
	<type>jalali</type>
	<year>1402</year>
	<month>6</month>
	<day>1</day>
</pubdate>
<pubdate>
	<type>gregorian</type>
	<year>2023</year>
	<month>9</month>
	<day>1</day>
</pubdate>
<volume>20</volume>
<number>2</number>
<publish_type>online</publish_type>
<publish_edition>1</publish_edition>
<article_type>fulltext</article_type>
<articleset>
	<article>


	<language>fa</language>
	<article_id_doi></article_id_doi>
	<title_fa>الگوسازی  موضوع‌ها بر پایه‌ی روش بیز گوناگونی</title_fa>
	<title>Topic Modeling Based on Variational Bayes Method</title>
	<subject_fa>مقالات پردازش متن </subject_fa>
	<subject>Paper</subject>
	<content_type_fa>كاربردي</content_type_fa>
	<content_type>Applicable</content_type>
	<abstract_fa>&lt;strong&gt;در این مقاله، برپایه&#8204;ی روش بیز گوناگونی، نشان می&#8204;دهیم که روش تخصیص پنهان دیریکله که یک مدل احتمالاتی مولّد است و در پردازش زبان&#8204;های طبیعی، متن&#8204;کاوی، کاهش ابعاد، و زیست&#8204;داده&#8204;ورزی کاربرد دارد، &lt;/strong&gt;&amp;nbsp;&lt;strong&gt;نسبت به روش تحلیل معنایی پنهان احتمالاتی در مدل&#8204;بندی داده&#8204;ها عملکرد بهتری دارد. در این باره، ابتدا یک مدل بیزی را در مدل&#8204;سازی موضوع&#8204;ها شرح می&#8204;دهیم. آنگاه با روش بیز گوناگونی و&lt;/strong&gt; &lt;strong&gt;الگوریتم &lt;/strong&gt;&lt;strong&gt;امیدریاضی-بیشینه&lt;/strong&gt;&lt;strong&gt;&lt;span dir=&quot;LTR&quot;&gt;&#8204;&lt;/span&gt;&lt;/strong&gt;&lt;strong&gt;سازی (&lt;/strong&gt;&lt;strong&gt;&lt;span dir=&quot;LTR&quot;&gt;EM&lt;/span&gt;&lt;/strong&gt;&lt;strong&gt;) پارامترهای مدل را برآورد می&#8204;کنیم. سپس الگوریتم ارائه شده، موسوم به الگوریتم &lt;/strong&gt;&lt;strong&gt;&lt;span dir=&quot;LTR&quot;&gt;EM&lt;/span&gt;&lt;/strong&gt; &lt;strong&gt;گوناگونی،&lt;/strong&gt;&lt;strong&gt; را برپایه&#8204;ی یک مجموعه&#8204;داده&#8204;ی نوشتاری از داده&#8204;های واقعی در زمینه&#8204;ی تحلیل داده&#8204;های خبری پیاده&#8204;سازی می&#8204;کنیم و مدل&#8204;بندی زبانی را بر اساس ملاک سرگشتگی بررسی می&#8204;کنیم، و دقت خوشه&#8204;بندی موضوع&#8204;ها و کاربرد کاهش ابعاد داده&#8204;های حجیم را با کمک ماشین بردار پشتیبان می&#8204;سنجیم. همچنین در مقایسه&#8204;ای دیگر، کاربرد الگوریتم پیشنهادی را در پالایش همکارانه بررسی می&#8204;کنیم.&lt;/strong&gt;&lt;strong&gt;&lt;span dir=&quot;LTR&quot;&gt;&lt;/span&gt;&lt;/strong&gt;</abstract_fa>
	<abstract>&lt;strong&gt;The Latent Dirichlet Allocation (LDA) model is a generative model with several applications in natural language processing, text mining, dimension reduction, and bioinformatics. It is a powerful technique in topic modeling in text mining, which is a data mining method to categorize documents by their topic.&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;Basic methods for topic modeling, including TF-IDF, unigram, and mixture of unigrams successfully deployed in modern search engines. Although these methods have some useful benefits, they don&amp;rsquo;t provide much summarization and reduction. To overcome these shortcomings, the latent semantic analysis (LSA) has been proposed, which uses singular value decomposition (SVD) of word-document matrix to compress big collection of text corpora. User&amp;rsquo;s search key words can be queried by making a pseudo-document vector. The next improvement step in topic modeling was probabilistic latent semantic analysis (PLSA), which has a close relation to LSA and matrix decomposition with SVD. By introducing of exchangeability for the words in documents, the topic modeling has been proceeded beyond PLSA and leads to LDA model.&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;We consider a corpus &lt;/strong&gt;&lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image001.png&quot; &gt; &lt;strong&gt;&amp;nbsp;contains M&lt;/strong&gt;&amp;nbsp;&lt;strong&gt;&amp;nbsp;documents, each document &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image003.png&quot; &gt; &lt;strong&gt;&amp;nbsp;has &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image004.png&quot; &gt; &lt;strong&gt;&amp;nbsp;words, and each word is an indicator from one of &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image005.png&quot; &gt; &lt;strong&gt;&amp;nbsp;vocabularies. We defined a generative model for generation of each document as follows. For each document draw its topic &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image006.png&quot; &gt; &lt;strong&gt;&amp;nbsp;from &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image007.png&quot; &gt; &lt;strong&gt;&amp;nbsp;and repeatedly for each &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image008.png&quot; &gt; &lt;strong&gt;&amp;nbsp;draw topic of each word &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image009.png&quot; &gt; &lt;strong&gt;&amp;nbsp;from &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image010.png&quot; &gt; &lt;strong&gt;&amp;nbsp;and draw each word from the probability matrix of&lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image011.png&quot; &gt; &lt;strong&gt;&amp;nbsp;with probability of &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image012.png&quot; &gt; &lt;strong&gt;. We can repeat this procedure to generate whole documents of corpus. We want to find corpus related parameters &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image013.png&quot; &gt; &lt;strong&gt;&amp;nbsp;and &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image014.png&quot; &gt; &lt;strong&gt;&amp;nbsp;as well as latent variables &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image015.png&quot; &gt; &lt;strong&gt;&amp;nbsp;and &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image006.png&quot; &gt; &lt;strong&gt;&amp;nbsp;for each document. Unfortunately, the posterior &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image016.png&quot; &gt; &lt;strong&gt;&amp;nbsp;is intractable, and we have to choose an approximation scheme.&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;In this paper we utilize LDA for collection of discrete text corpora. We describe procedures for inference and parameter estimation. Since computing posterior distribution of hidden variables given a document is intractable to compute in general, we use approximate inference algorithm called variational Bayes method. The basic idea of variational Bayes is to consider a family of adjustable lower bound on the posterior, then finds the tightest possible one. To estimate optimal hyper-parameters in the model, we used the empirical Bayes method, as well as a specialized expectation-maximization (EM) algorithm called variational-EM algorithm.&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;The results are reported in document modeling, text classification, and collaborative filtering. The topic modeling of LDA and PLSA models are compared on a Persian news data set. It has been observed that LDA has perplexity between &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image017.png&quot; &gt; &lt;strong&gt;&amp;nbsp;and &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image018.png&quot; &gt; &lt;strong&gt;, while the PLSA has perplexity between &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image019.png&quot; &gt; &lt;strong&gt;&amp;nbsp;and &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image020.png&quot; &gt; &lt;strong&gt;, which shows domination of LDA over PLSA.&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;The LDA model has also been applied for dimension reduction in a document classification problem, along with the support vector machines (SVM) classification method. Two competitor models are compared, first trained on a low-dimensional representation provided by LDA and the second trained on all documents of corpus, with accuracies &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image021.png&quot; &gt; &lt;strong&gt;&amp;nbsp;and &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image022.png&quot; &gt; &lt;strong&gt;, respectively, this means we lose accuracy but it remains in reasonable range when we use LDA model for dimensionality reduction.&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;Finally, we used the LDA and PLSA methods along with the collaborative filtering for MovieLens 1m data set, and we observed that the predictive-perplexity of LDA changes from &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image023.png&quot; &gt; &lt;strong&gt;&amp;nbsp;to &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image024.png&quot; &gt; &lt;strong&gt;&amp;nbsp;while it changes from &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image025.png&quot; &gt; &lt;strong&gt;&amp;nbsp;to &lt;/strong&gt; &lt;img alt=&quot;&quot; chromakey=&quot;white&quot; src=&quot;file:///C:/Users/SF1A4~1.MTA/AppData/Local/Temp/msohtmlclip1/01/clip_image026.png&quot; &gt; &lt;strong&gt;&amp;nbsp;for PLSA, again showing the domination of the LDA method.&lt;/strong&gt;&lt;strong&gt;&lt;span dir=&quot;RTL&quot;&gt;&lt;/span&gt;&lt;/strong&gt;</abstract>
	<keyword_fa>روش بیز گوناگونی, تخصیص پنهان دیریکله, الگوریتم امیدریاضی-بیشینه‌سازی, یادگیری ماشین, پردازش زبان‌های طبیعی</keyword_fa>
	<keyword>Variational Bayes method, Latent Dirichlet allocation, Expectation-Maximization algorithm, Machine learning, Natural language processing</keyword>
	<start_page>39</start_page>
	<end_page>58</end_page>
	<web_url>http://jsdp.rcisp.ac.ir/browse.php?a_code=A-10-2200-1&amp;slc_lang=fa&amp;sid=1</web_url>


<author_list>
	<author>
	<first_name>Vahid</first_name>
	<middle_name></middle_name>
	<last_name>Heidari</last_name>
	<suffix></suffix>
	<first_name_fa>وحید</first_name_fa>
	<middle_name_fa></middle_name_fa>
	<last_name_fa>حیدری</last_name_fa>
	<suffix_fa></suffix_fa>
	<email>vahid.heidari@ut.ac.ir</email>
	<code>100319475328460012427</code>
	<orcid>100319475328460012427</orcid>
	<coreauthor>No</coreauthor>
	<affiliation></affiliation>
	<affiliation_fa>دانشگاه تهران</affiliation_fa>
	 </author>


	<author>
	<first_name>S. Mahmoud</first_name>
	<middle_name></middle_name>
	<last_name>Taheri</last_name>
	<suffix></suffix>
	<first_name_fa>سید محمود</first_name_fa>
	<middle_name_fa></middle_name_fa>
	<last_name_fa>طاهری</last_name_fa>
	<suffix_fa></suffix_fa>
	<email>sm_taheri@ut.ac.ir</email>
	<code>100319475328460012428</code>
	<orcid>100319475328460012428</orcid>
	<coreauthor>Yes
</coreauthor>
	<affiliation></affiliation>
	<affiliation_fa>دانشگاه تهران</affiliation_fa>
	 </author>


	<author>
	<first_name>Morteza</first_name>
	<middle_name></middle_name>
	<last_name>Amini</last_name>
	<suffix></suffix>
	<first_name_fa>مرتضی</first_name_fa>
	<middle_name_fa></middle_name_fa>
	<last_name_fa>امینی</last_name_fa>
	<suffix_fa></suffix_fa>
	<email>morteza.amini@ut.ac.ir</email>
	<code>100319475328460012429</code>
	<orcid>100319475328460012429</orcid>
	<coreauthor>No</coreauthor>
	<affiliation></affiliation>
	<affiliation_fa>دانشگاه تهران</affiliation_fa>
	 </author>


</author_list>


	</article>
</articleset>
</journal>
