{"id":1021,"date":"2022-11-02T17:55:39","date_gmt":"2022-11-02T16:55:39","guid":{"rendered":"https:\/\/www.nard-intelligence.net\/?p=1021"},"modified":"2023-05-15T23:59:16","modified_gmt":"2023-05-15T21:59:16","slug":"hui2vec-learning-transaction-embedding-through-high-utility-itemsets-2","status":"publish","type":"post","link":"https:\/\/www.nard-intelligence.net\/fr\/blog\/hui2vec-learning-transaction-embedding-through-high-utility-itemsets-2\/","title":{"rendered":"Hui2Vec: Learning Transaction Embedding Through High Utility Itemsets"},"content":{"rendered":"\n<p><\/p>\n\n\n\n<p>NARD Intelligence est ravie d&rsquo;annoncer qu&rsquo;un projet de recherche en collaboration entre des chercheurs du d\u00e9partement Informatique de l&rsquo;<a href=\"http:\/\/www.istic.rnu.tn\/fr\/accueil.html?catid=0&amp;id=115\">Institut Sup\u00e9rieur des Technologies de l&rsquo;Information<\/a> (Tunisie) et de la Communication (Tunisie) et du Big Data Institute de <a href=\"https:\/\/en.szu.edu.cn\/\">l&rsquo;Universit\u00e9 de Shenzhen<\/a> (Chine) a \u00e9t\u00e9 accept\u00e9 pour publication dans la conf\u00e9rence&nbsp;<a href=\"https:\/\/www.bda2022.org\/\">International Conference on Big Data Analytics (BDA 2022)<\/a>&nbsp;qui se tiendra du 19 au 22 d\u00e9cembre 2022 \u00e0 Hederabad, en Inde.<\/p>\n\n\n\n<p>Liste de tous les papiers accept\u00e9es \u00e0 la conf\u00e9rence : <a href=\"https:\/\/www.bda2022.org\/Programme\/AcceptedPapers\">BDA Accepted Papers List<\/a> <\/p>\n\n\n\n<p>Ce travail pr\u00e9sente une nouvelle approche pour l&rsquo;apprentissage des plongements (embeddings) des tranactions sur la base des &lsquo;Itemsets&rsquo; \u00e0 haute valeur utile. Il s&rsquo;agit d&rsquo;une adaptation de <a href=\"https:\/\/en.wikipedia.org\/wiki\/Word2vec\">Word2vec<\/a>, la tr\u00e8s c\u00e9l\u00e8bre technique de <a href=\"https:\/\/en.wikipedia.org\/wiki\/Natural_language_processing\">Natural Language Processing<\/a> (NLP) qui utilise un mod\u00e8le de r\u00e9seau de neurones pour apprendre les plongements de mots (word embeddings) \u00e0 partir d&rsquo;un grand corpus de texte, au domaine  pattern\/itemset mining.<\/p>\n\n\n\n<h4 class=\"wp-block-heading\">Abstract<\/h4>\n\n\n\n<p>Mining frequent itemsets (FIs) in transaction databases is a very popular task in data mining. It helps create meaningful and effective representations for customer transactions which is a key step in the process of transaction classification and clustering. To improve the quality of these representations, previous studies have adapted vector embedding methods to learn transaction embeddings from items and FIs.<br>However, FIs are still a simple pattern type that ignores important information about transactions such as the purchase quantities of items and their unit profits.<br>To address this issue, we propose to learn transaction embeddings from items and high-utility itemsets (HUIs), a more general pattern type. Since HUIs were shown to be more appropriate than FIs for a wide range of applications, we take for hypothesis that transaction embeddings learned from HUIs will be more representative and meaningful. We introduce an unsupervised method, named Hui2Vec, to learn  transaction embeddings by combining both singleton items and HUIs. We demonstrate the superior quality of the embedding achieved with the proposed method compared to the embeddings learned from items and FIs on four datasets.<\/p>\n\n\n\n<p>#word-embedding #machine-learning #ml #ai #high-utility #data-mining <\/p>\n","protected":false},"excerpt":{"rendered":"<p>NARD Intelligence est ravie d&rsquo;annoncer qu&rsquo;un projet de recherche en collaboration entre des chercheurs du d\u00e9partement Informatique de l&rsquo;Institut Sup\u00e9rieur &#8230; <a class=\"cz_readmore\" href=\"https:\/\/www.nard-intelligence.net\/fr\/blog\/hui2vec-learning-transaction-embedding-through-high-utility-itemsets-2\/\"><i class=\"fa fa-angle-right\"><\/i><span>More<\/span><\/a><\/p>\n","protected":false},"author":3,"featured_media":1016,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[26,28],"tags":[],"yst_prominent_words":[],"class_list":["post-1021","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-fr","category-research-fr"],"_links":{"self":[{"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/posts\/1021","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/users\/3"}],"replies":[{"embeddable":true,"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/comments?post=1021"}],"version-history":[{"count":4,"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/posts\/1021\/revisions"}],"predecessor-version":[{"id":1139,"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/posts\/1021\/revisions\/1139"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/media\/1016"}],"wp:attachment":[{"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/media?parent=1021"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/categories?post=1021"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/tags?post=1021"},{"taxonomy":"yst_prominent_words","embeddable":true,"href":"https:\/\/www.nard-intelligence.net\/fr\/wp-json\/wp\/v2\/yst_prominent_words?post=1021"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}