1
|
Shafiq N, Hamid I, Asif M, Nawaz Q, Aljuaid H, Ali H. Abstractive text summarization of low-resourced languages using deep learning. PeerJ Comput Sci 2023; 9:e1176. [PMID: 37346684 PMCID: PMC10280265 DOI: 10.7717/peerj-cs.1176] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2022] [Accepted: 11/09/2022] [Indexed: 06/23/2023]
Abstract
Background Humans must be able to cope with the huge amounts of information produced by the information technology revolution. As a result, automatic text summarization is being employed in a range of industries to assist individuals in identifying the most important information. For text summarization, two approaches are mainly considered: text summarization by the extractive and abstractive methods. The extractive summarisation approach selects chunks of sentences like source documents, while the abstractive approach can generate a summary based on mined keywords. For low-resourced languages, e.g., Urdu, extractive summarization uses various models and algorithms. However, the study of abstractive summarization in Urdu is still a challenging task. Because there are so many literary works in Urdu, producing abstractive summaries demands extensive research. Methodology This article proposed a deep learning model for the Urdu language by using the Urdu 1 Million news dataset and compared its performance with the two widely used methods based on machine learning, such as support vector machine (SVM) and logistic regression (LR). The results show that the suggested deep learning model performs better than the other two approaches. The summaries produced by extractive summaries are processed using the encoder-decoder paradigm to create an abstractive summary. Results With the help of Urdu language specialists, the system-generated summaries were validated, showing the proposed model's improvement and accuracy.
Collapse
Affiliation(s)
- Nida Shafiq
- Department of Computer Science, National Textile University, Faisalabad, Pakistan
| | - Isma Hamid
- Department of Computer Science, National Textile University, Faisalabad, Pakistan
| | - Muhammad Asif
- Department of Computer Science, National Textile University, Faisalabad, Pakistan
| | - Qamar Nawaz
- Department of Computer Science, University of Agriculture Faisalabad, Faisalabad, Pakistan
| | - Hanan Aljuaid
- Computer Sciences Department, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia
| | - Hamid Ali
- Department of Computer Science, National Textile University, Faisalabad, Pakistan
| |
Collapse
|
2
|
An Examination of Classical Art Impact and Popularity through Social Media Emotion Analysis of Art Memes and Museum Posts. INFORMATION 2022. [DOI: 10.3390/info13100468] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022] Open
Abstract
On Instagram, we have all seen memes. Honestly, what would you do if you encountered a meme in a museum? The purpose of the study is to evaluate the nexus between posts uploaded by museum visitors and emotions, as well as the popularity of artworks and memes. We gathered N = 4.526 (N = 1.222 for memes and N = 3.304 for museum posts) entire posts using API. We selected the total number of likes, comments, frequency, nwords, and text emotions as indicators for several supervised machine learning tasks. Moreover, we used a ranking algorithm to measure meme and artwork popularity. Our experiments revealed the most prevalent emotions in both the memes dataset and museum posts dataset. The ranking task showed the most popular meme and museum post, respectively, that can influence the aesthetic experience and its popularity. This study provided further insight into the social media sphere that has had a significant effect on the aesthetic experience of museums and artwork’s popularity. As a final point, we anticipate that our outcomes will serve as a springboard for future studies in social media, art, and cultural analytics.
Collapse
|
3
|
Huang Y, Huang J, Chen X, Wang Q, Meng H. An end-to-end heterogeneous network for graph similarity learning. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.07.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
4
|
Hu Y, Wen G, Luo M, Dai D, Cao W, Yu Z, Hall W. Inner-Imaging Networks: Put Lenses Into Convolutional Structure. IEEE TRANSACTIONS ON CYBERNETICS 2022; 52:8547-8560. [PMID: 34398768 DOI: 10.1109/tcyb.2020.3034605] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Despite the tremendous success in computer vision, deep convolutional networks suffer from serious computation costs and redundancies. Although previous works address that by enhancing the diversities of filters, they have not considered the complementarity and the completeness of the internal convolutional structure. To respond to this problem, we propose a novel inner-imaging (InI) architecture, which allows relationships between channels to meet the above requirement. Specifically, we organize the channel signal points in groups using convolutional kernels to model both the intragroup and intergroup relationships simultaneously. A convolutional filter is a powerful tool for modeling spatial relations and organizing grouped signals, so the proposed methods map the channel signals onto a pseudoimage, like putting a lens into the internal convolution structure. Consequently, not only is the diversity of channels increased but also the complementarity and completeness can be explicitly enhanced. The proposed architecture is lightweight and easy to be implement. It provides an efficient self-organization strategy for convolutional networks to improve their performance. Extensive experiments are conducted on multiple benchmark datasets, including CIFAR, SVHN, and ImageNet. Experimental results verify the effectiveness of the InI mechanism with the most popular convolutional networks as the backbones.
Collapse
|
5
|
Jarrahi A, Safari L. Evaluating the effectiveness of publishers' features in fake news detection on social media. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 82:2913-2939. [PMID: 35431607 PMCID: PMC8995145 DOI: 10.1007/s11042-022-12668-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/20/2021] [Revised: 12/21/2021] [Accepted: 02/21/2022] [Indexed: 06/14/2023]
Abstract
With the expansion of the Internet and attractive social media infrastructures, people prefer to follow the news through these media. Despite the many advantages of these media in the news field, the lack of control and verification mechanism has led to the spread of fake news as one of the most critical threats to democracy, economy, journalism, health, and freedom of expression. So, designing and using efficient automated methods to detect fake news on social media has become a significant challenge. One of the most relevant entities in determining the authenticity of a news statement on social media is its publishers. This paper examines the publishers' features in detecting fake news on social media, including Credibility, Influence, Sociality, Validity, and Lifetime. In this regard, we propose an algorithm, namely CreditRank, for evaluating publishers' credibility on social networks. We also suggest a high accurate multi-modal framework, namely FR-Detect, for fake news detection using user-related and content-related features. Furthermore, a sentence-level convolutional neural network is provided to properly combine publishers' features with latent textual content features. Experimental results show that the publishers' features can improve the performance of content-based models by up to 16% and 31% in accuracy and F1, respectively. Also, the behavior of publishers in different news domains has been statistically studied and analyzed.
Collapse
Affiliation(s)
- Ali Jarrahi
- Electrical and Computer Engineering, University of Zanjan, Zanjan, Iran
| | - Leila Safari
- Electrical and Computer Engineering, University of Zanjan, Zanjan, Iran
| |
Collapse
|
6
|
Antunes N, Ferreira JC, Cardoso E. Generating personalized business card designs from images. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 81:25051-25073. [PMID: 35342325 PMCID: PMC8939401 DOI: 10.1007/s11042-022-12416-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/02/2020] [Revised: 04/04/2021] [Accepted: 01/25/2022] [Indexed: 06/14/2023]
Abstract
Rising competition in the retail and hospitality sectors, especially in densely populated and touristic destinations is a growing concern for many business owners, who wish to deliver their brand communication strategy to the target audience. Many of these businesses rely on word-of-mouth marketing, delivering business cards to customers. Furthermore, the lack of a dedicated marketing team and budget for brand image consolidation and design creation often limits the brand expansion capability. The purpose of this study is to propose a novel system prototype that can suggest personalized designs for business cards, based on an existing business card picture. Using perspective transformation, text extraction and colour reduction techniques, we were able to obtain features from the original business card image and generate an alternative design, personalized for the end user. We have successfully been able to generate customized business cards for different business types, with textual information and a custom colour palette matching the original submitted image. All of the system modules were demonstrated to have positive results for the test cases and the proposal answered the main research question. Further research and development is required to adapt the current system to other marketing printouts, such as flyers or posters.
Collapse
Affiliation(s)
- Nuno Antunes
- ISTAR, Instituto Universitário de Lisboa (ISCTE-IUL), 1649-026 Lisboa, Portugal
- INOV Instituto de Engenharia de Sistemas e Computadores Inovação, Rua Alves Redol, 9, 1000-029 Lisbon, Portugal
| | - João Carlos Ferreira
- ISTAR, Instituto Universitário de Lisboa (ISCTE-IUL), 1649-026 Lisboa, Portugal
- INOV Instituto de Engenharia de Sistemas e Computadores Inovação, Rua Alves Redol, 9, 1000-029 Lisbon, Portugal
| | - Elsa Cardoso
- ISTAR, Instituto Universitário de Lisboa (ISCTE-IUL), 1649-026 Lisboa, Portugal
| |
Collapse
|
7
|
Suleiman D, Awajan A. Multilayer encoder and single-layer decoder for abstractive Arabic text summarization. Knowl Based Syst 2022. [DOI: 10.1016/j.knosys.2021.107791] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
8
|
|
9
|
Ye H, Li H, Chen CLP. Adaptive Deep Cascade Broad Learning System and Its Application in Image Denoising. IEEE TRANSACTIONS ON CYBERNETICS 2021; 51:4450-4463. [PMID: 32203051 DOI: 10.1109/tcyb.2020.2978500] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
This article proposes a novel regularization deep cascade broad learning system (DCBLS) architecture, which includes one cascaded feature mapping nodes layer and one cascaded enhancement nodes layer. Then, the transformation feature representation is easily obtained by incorporating the enhancement nodes and the feature mapping nodes. Once such a representation is established, a final output layer is constructed by implementing a simple convex optimization model. Furthermore, a parallelization framework on the new method is designed to make it compatible with large-scale data. Simultaneously, an adaptive regularization parameter criterion is adopted under some conditions. Moreover, the stability and error estimate of this method are discussed and proved mathematically. The proposed method could extract sufficient available information from the raw data compared with the standard broad learning system and could achieve compellent successes in image denoising. The experiments results on benchmark datasets, including natural images as well as hyperspectral images, verify the effectiveness and superiority of the proposed method in comparison with the state-of-the-art approaches for image denoising.
Collapse
|
10
|
Shen J, Robertson N. BBAS: Towards large scale effective ensemble adversarial attacks against deep neural network learning. Inf Sci (N Y) 2021. [DOI: 10.1016/j.ins.2020.11.026] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
11
|
Dun Y, Da Z, Yang S, Qian X. Image super-resolution based on residually dense distilled attention network. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.02.008] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
12
|
Liu D, Liang C, Chen S, Tie Y, Qi L. Auto-encoder based structured dictionary learning for visual classification. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2020.09.088] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
13
|
Chen H, Hu C, Lee F, Lin C, Yao W, Chen L, Chen Q. A Supervised Video Hashing Method Based on a Deep 3D Convolutional Neural Network for Large-Scale Video Retrieval. SENSORS 2021; 21:s21093094. [PMID: 33946745 PMCID: PMC8124307 DOI: 10.3390/s21093094] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/15/2021] [Revised: 04/20/2021] [Accepted: 04/26/2021] [Indexed: 11/24/2022]
Abstract
Recently, with the popularization of camera tools such as mobile phones and the rise of various short video platforms, a lot of videos are being uploaded to the Internet at all times, for which a video retrieval system with fast retrieval speed and high precision is very necessary. Therefore, content-based video retrieval (CBVR) has aroused the interest of many researchers. A typical CBVR system mainly contains the following two essential parts: video feature extraction and similarity comparison. Feature extraction of video is very challenging, previous video retrieval methods are mostly based on extracting features from single video frames, while resulting the loss of temporal information in the videos. Hashing methods are extensively used in multimedia information retrieval due to its retrieval efficiency, but most of them are currently only applied to image retrieval. In order to solve these problems in video retrieval, we build an end-to-end framework called deep supervised video hashing (DSVH), which employs a 3D convolutional neural network (CNN) to obtain spatial-temporal features of videos, then train a set of hash functions by supervised hashing to transfer the video features into binary space and get the compact binary codes of videos. Finally, we use triplet loss for network training. We conduct a lot of experiments on three public video datasets UCF-101, JHMDB and HMDB-51, and the results show that the proposed method has advantages over many state-of-the-art video retrieval methods. Compared with the DVH method, the mAP value of UCF-101 dataset is improved by 9.3%, and the minimum improvement on JHMDB dataset is also increased by 0.3%. At the same time, we also demonstrate the stability of the algorithm in the HMDB-51 dataset.
Collapse
Affiliation(s)
- Hanqing Chen
- Shanghai Engineering Research Center of Assistive Devices, School of Medical Instrument and Food Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China; (H.C.); (C.L.); (W.Y.); (L.C.)
| | - Chunyan Hu
- School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China;
| | - Feifei Lee
- Shanghai Engineering Research Center of Assistive Devices, School of Medical Instrument and Food Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China; (H.C.); (C.L.); (W.Y.); (L.C.)
- Correspondence: (F.L.); (Q.C.)
| | - Chaowei Lin
- Shanghai Engineering Research Center of Assistive Devices, School of Medical Instrument and Food Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China; (H.C.); (C.L.); (W.Y.); (L.C.)
| | - Wei Yao
- Shanghai Engineering Research Center of Assistive Devices, School of Medical Instrument and Food Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China; (H.C.); (C.L.); (W.Y.); (L.C.)
| | - Lu Chen
- Shanghai Engineering Research Center of Assistive Devices, School of Medical Instrument and Food Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China; (H.C.); (C.L.); (W.Y.); (L.C.)
| | - Qiu Chen
- Major of Electrical Engineering and Electronics, Graduate School of Engineering, Kogakuin University, Tokyo 163-8677, Japan
- Correspondence: (F.L.); (Q.C.)
| |
Collapse
|
14
|
Xue Y, Li Y, Liu S, Zhang X, Qian X. Crowd Scene Analysis Encounters High Density and Scale Variation. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:2745-2757. [PMID: 33502976 DOI: 10.1109/tip.2021.3049963] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Crowd scene analysis receives growing attention due to its wide applications. Grasping the accurate crowd location is important for identifying high-risk regions. In this article, we propose a Compressed Sensing based Output Encoding (CSOE) scheme, which casts detecting pixel coordinates of small objects into a task of signal regression in encoding signal space. To prevent gradient vanishing, we derive our own sparse reconstruction backpropagation rule that is adaptive to distinct implementations of sparse reconstruction and makes the whole model end-to-end trainable. With the support of CSOE and the backpropagation rule, the proposed method shows more robustness to deep model training error, which is especially harmful to crowd counting and localization. The proposed method achieves state-of-the-art performance across four mainstream datasets, especially achieves excellent results in highly crowded scenes. A series of analysis and experiments support our claim that regression in CSOE space is better than traditionally detecting coordinates of small objects in pixel space for highly crowded scenes.
Collapse
|
15
|
Wang Y, Wei Y, Qian X, Zhu L, Yang Y. Sketch-Guided Scenery Image Outpainting. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:2643-2655. [PMID: 33523812 DOI: 10.1109/tip.2021.3054477] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The outpainting results produced by existing approaches are often too random to meet users' requirements. In this work, we take the image outpainting one step forward by allowing users to harvest personal custom outpainting results using sketches as the guidance. To this end, we propose an encoder-decoder based network to conduct sketch-guided outpainting, where two alignment modules are adopted to impose the generated content to be realistic and consistent with the provided sketches. First, we apply a holistic alignment module to make the synthesized part be similar to the real one from the global view. Second, we reversely produce the sketches from the synthesized part and encourage them be consistent with the ground-truth ones using a sketch alignment module. In this way, the learned generator will be imposed to pay more attention to fine details and be sensitive to the guiding sketches. To our knowledge, this work is the first attempt to explore the challenging yet meaningful conditional scenery image outpainting. We conduct extensive experiments on two collected benchmarks to qualitatively and quantitatively validate the effectiveness of our approach compared with the other state-of-the-art generative models.
Collapse
|
16
|
|
17
|
Xiang J, Zhang N, Pan R, Gao W. Fabric Retrieval Based on Multi-Task Learning. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:1570-1582. [PMID: 33373301 DOI: 10.1109/tip.2020.3043877] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Due to the potential values in many areas such as e-commerce and inventory management, fabric image retrieval, which is a special case in Content Based Image Retrieval (CBIR), has recently become a research hotspot. It is also a challenging issue with serval obstacles: variety and complexity of fabric appearance, high requirements for retrieval accuracy. To address this issue, this paper proposes a novel approach for fabric image retrieval based on multi-task learning and deep hashing. According to the cognitive system of fabric, a multi-classification-task learning model with uncertainty loss and constraint is presented to learn fabric image representation. Then we adopt an unsupervised deep network to encode the extracted features into 128-bits hashing codes. Further, the hashing codes are regarded as the index of fabrics image for image retrieval. To evaluate the proposed approach, we expanded and upgraded the dataset WFID, which was built in our previous research specifically for fabric image retrieval. The experimental results show that the proposed approach outperforms the state-of-the-art.
Collapse
|
18
|
Zhang B, Ling H, Shen J, Wang Q, Lei J, Shi Y, Wu L, Li P. Mixture Distribution Graph Network for Few Shot Learning. IEEE Trans Cogn Dev Syst 2021. [DOI: 10.1109/tcds.2021.3075280] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|