Read by QxMD icon Read

image captioning

Etienne Xavier Keller, Steeve Doizi, Luca Villa, Olivier Traxer
PURPOSE: To present attributes of currently available flexible ureteroscopes to define the best flexible ureteroscope for upper tract urothelial carcinoma (UTUC) treatment. MATERIALS AND METHODS: Scopus and Medline databases were searched for articles relating to performance of flexible ureteroscopes. A consensus for final inclusion of articles judged to be relevant for UTUC treatment was reached between the authors. Instrument characteristics were extracted from manufacturers' product brochures...
February 15, 2019: World Journal of Urology
Lianli Gao, Xiangpeng Li, Jingkuan Song, Heng Tao Shen
Recent progress has been made in using attention based encoder-decoder framework for image and video captioning. Most existing decoders apply the attention mechanism to every generated word including both visual words (e.g., "gun" and "shooting") and non-visual words (e.g. "the", "a"). However, these non-visual words can be easily predicted using natural language model without considering visual signals or attention. Imposing attention mechanism on non-visual words could mislead and decrease the overall performance of visual captioning...
January 21, 2019: IEEE Transactions on Pattern Analysis and Machine Intelligence
Man Zhang, Li-Li Huang, Chen-Huai Teng, Fang-Fang Wu, Li-Yun Ge, Yu-Juan Shi, Zheng-Le He, Lei Liu, Cheng-Jie Jiang, Ruo-Nan Hou, Jian Xiao, Hong-Yu Zhang, Da-Qing Chen
The original version of this article unfortunately contained a mistake. The Fluorescence Immunoassays text written in Materials and Methods section and Fig. 1i, j is incorrect. In Fig. 1j, the images corresponding to Sham and TBI + ILG are incorrect. In Fig. 1i the figure caption "TBI + EDA" are incorrect. The corrected text and Fig. 1i, j are given below.
January 12, 2019: Neurochemical Research
Niange Yu, Xiaolin Hu, Binheng Song, Jian Yang, Jianwei Zhang
We present an image captioning framework that generates captions under a given topic. The topic candidates are extracted from the caption corpus. A given image's topics are then selected from these candidates by a CNN-based multilabel classifier. The input to the caption generation model is an image-topic pair, and the output is a caption of the image. For this purpose, a cross-modal embedding method is learned for the images, topics, and captions. In the proposed framework, the topic, caption, and image are organized in a hierarchical structure which is preserved in the embedding space by using the orderembedding method...
December 27, 2018: IEEE Transactions on Image Processing: a Publication of the IEEE Signal Processing Society
Mara Mariconti, Ambra Vola, Tommaso Manciulli, Francesca Genco, Raffaella Lissandrin, Valeria Meroni, Mara Rosenzvit, Francesca Tamarozzi, Enrico Brunetti
The published article unfortunately contained mistake. After replacing the image from colored version to black and white, the Fig. 1 caption, unfortunately, was left unupdated.
December 27, 2018: Immunologic Research
Feiran Huang, Xiaoming Zhang, Zhoujun Li, Zhonghua Zhao
Image-text matching by deep models has recently made remarkable achievements in many tasks, such as image caption and image search. A major challenge of matching the image and text lies in that they usually have complicated underlying relations between them and simply modeling the relations may lead to suboptimal performance. In this paper, we develop a novel approach Bi-directional Spatial-Semantic Attention Networks (BSSAN), which leverages both the word to regions (W2R) relation and image object to words (O2W) relation in a holistic deep framework for more effectively matching...
November 19, 2018: IEEE Transactions on Image Processing: a Publication of the IEEE Signal Processing Society
Konda Reddy Mopuri, Utsav Garg, R Venkatesh Babu
Deep convolutional neural networks (CNN) have revolutionized various fields of vision research and have seen unprecedented adoption for multiple tasks such as classification, detection, captioning, etc. However, they offer little transparency into their inner workings and are often treated as black boxes that deliver excellent performance. In this work, we aim at alleviating this opaqueness of CNNs by providing visual explanations for the network's predictions. Our approach can analyze a variety of CNN based models trained for vision applications such as object recognition and caption generation...
November 16, 2018: IEEE Transactions on Image Processing: a Publication of the IEEE Signal Processing Society
Tharindu Fernando, Simon Denman, Sridha Sridharan, Clinton Fookes
As humans we possess an intuitive ability for navigation which we master through years of practice; however existing approaches to model this trait for diverse tasks including monitoring pedestrian flow and detecting abnormal events have been limited by using a variety of hand-crafted features. Recent research in the area of deep-learning has demonstrated the power of learning features directly from the data; and related research in recurrent neural networks has shown exemplary results in sequence-to-sequence problems such as neural machine translation and neural image caption generation...
September 20, 2018: Neural Networks: the Official Journal of the International Neural Network Society
Harry A Atwater, Artur R Davoyan, Ognjen Ilic, Deep Jariwala, Michelle C Sherrott, Cora M Went, William S Whitney, Joeson Wong
In the version of this Perspective originally published, Fig. 1 was missing the following credit line from the caption: 'Background image from ESA/Hubble (A. Fujii).' This has now been corrected in the online versions of the Perspective.
October 12, 2018: Nature Materials
Marcus A Badgeley, Manway Liu, Benjamin S Glicksberg, Mark Shervey, John Zech, Khader Shameer, Joseph Lehar, Eric K Oermann, Michael V McConnell, Thomas M Snyder, Joel T Dudley
Motivation: Radiologists have used algorithms for Computer-Aided Diagnosis (CAD) for decades. These algorithms use machine learning with engineered features, and there have been mixed findings on whether they improve radiologists' interpretations. Deep learning offers superior performance, but requires more training data and has not been evaluated in joint algorithm-radiologist decision systems. Results: We developed the Computer-Aided Note and Diagnosis Interface (CANDI) for collaboratively annotating radiographs and evaluating how algorithms alter human interpretation...
October 10, 2018: Bioinformatics
Mattia Gentile, Emanuele Agolini, Dario Cocciadiferro, Romina Ficarella, Emanuela Ponzi, Emanuele Bellacchio, Maria F Antonucci, Antonio Novelli
Biallelic exostosin-2 (EXT2) pathogenic variants have been described as the cause of the Seizures-Scoliosis-Macrocephaly syndrome (OMIM 616682) characterized by intellectual disability, facial dysmorphisms and seizures. More recently, it has been proposed to rename this disorder with the acronym AREXT2 (autosomal recessive EXT2-related syndrome). Here, we report the third family affected by AREXT2 syndrome, harboring compound missense variants in EXT2, p.Asp227Asn, and p.Tyr608Cys. In addition, our patients developed multiple exostoses, which were not observed in the previously described families...
October 4, 2018: Clinical Genetics
Maryam Ghanbarian, Mohammad Hossein Nicknam, Alireza Mesdaghinia, Masud Yunesian, Mohammad Sadegh Hassanvand, Narjes Soleimanifar, Soheila Rezaei, Zahra Atafar, Marjan Ghanbarian, Maryam Faraji, Mohammad Ghanbari Ghozikali, Kazem Naddafi
The original version of this article unfortunately contained a mistake. Figure 6 caption should be "The light microscopic image (a) and transmission electron microscopic image (b) of A549 cell after 24 h of exposure to PM10 (150 μg/ml).
September 4, 2018: Biological Trace Element Research
Pierre Antherieu, R Levy, T De Saint Denis, L Lohkamp, G Paternoster, F Di Rocco, N Boddaert, M Zerah
The article which was recently published contained error. The figures and figure captions were interchanged during the publication process of the paper.
August 22, 2018: Child's Nervous System: ChNS: Official Journal of the International Society for Pediatric Neurosurgery
Mingxing Zhang, Yang Yang, Hanwang Zhang, Yanli Ji, Heng Tao Shen, Tat-Seng Chua
Recently, a great progress in automatic image captioning has been achieved by using semantic concepts detected from the image. However, we argue that existing concepts-to-caption framework, in which the concept detector is trained using the image-caption pairs to minimize the vocabulary discrepancy, suffers from the deficiency of insufficient concepts. The reasons are two-fold: 1) the extreme imbalance between the number of occurrence positive and negative samples of the concept and 2) the incomplete labeling in training captions caused by the biased annotation and usage of synonyms...
January 2019: IEEE Transactions on Image Processing: a Publication of the IEEE Signal Processing Society
Senmao Ye, Nian Liu, Junwei Han
We propose a novel attention framework called attentive linear transformation (ALT). Instead of learning the spatial or channel-wise attention in existing models, ALT learns to attend to the high-dimensional transformation matrix from the image feature space to the context vector space. Thus ALT can learn various relevant feature abstractions, including spatial attention, channel-wise attention and visual dependence. Besides, we propose a soft threshold regression to predict the attention probabilities for local regions...
July 12, 2018: IEEE Transactions on Image Processing: a Publication of the IEEE Signal Processing Society
Cesc Chunseong Park, Byeongchang Kim, Gunhee Kim
We address personalized image captioning, which generates a descriptive sentence for a user's image, accounting for prior knowledge such as her active vocabularies or writing style in her previous documents. As applications of personalized image captioning, we solve two post automation tasks in social networks: hashtag prediction and post generation. The hashtag prediction predicts a list of hashtags for an image, while the post generation creates a natural post text consisting of normal words, emojis, and even hashtags...
April 10, 2018: IEEE Transactions on Pattern Analysis and Machine Intelligence
Xiao Xie, Xiwen Cai, Junpei Zhou, Nan Cao, Yingcai Wu
Interactive visualization of large image collections is important and useful in many applications, such as personal album management and user profiling on images. However, most prior studies focus on using low-level visual features of images, such as texture and color histogram, to create visualizations without considering the more important semantic information embedded in images. This paper proposes a novel visual analytic system to analyze images in a semantic-aware manner. The system mainly comprises two components: a semantic information extractor and a visual layout generator...
May 15, 2018: IEEE Transactions on Visualization and Computer Graphics
Kun Fu, Jin Li, Junqi Jin, Changshui Zhang
Image captioning aims to generate natural language sentences to describe the salient parts of a given image. Although neural networks have recently achieved promising results, a key problem is that they can only describe concepts seen in the training image-sentence pairs. Efficient learning of novel concepts has thus been a topic of recent interest to alleviate the expensive manpower of labeling data. In this paper, we propose a novel method, Image-Text Surgery, to synthesize pseudoimage-sentence pairs. The pseudopairs are generated under the guidance of a knowledge base, with syntax from a seed data set (i...
December 2018: IEEE Transactions on Neural Networks and Learning Systems
Sachin Muralidhara, Michael J Paul
BACKGROUND: Social media provides a complementary source of information for public health surveillance. The dominate data source for this type of monitoring is the microblogging platform Twitter, which is convenient due to the free availability of public data. Less is known about the utility of other social media platforms, despite their popularity. OBJECTIVE: This work aims to characterize the health topics that are prominently discussed in the image-sharing platform Instagram, as a step toward understanding how this data might be used for public health research...
June 29, 2018: JMIR Public Health and Surveillance
M Kaiser, M Jacobson, P H Andersen, P Bækbo, J J Cerón, J Dahl, D Escribano, S Jacobsen
The original article [1] contains an error whereby the caption in Figure 8 is incorrect; the correct caption can be seen ahead alongside its respective image.
June 1, 2018: BMC Veterinary Research
Fetch more papers »
Fetching more papers... Fetching...
Read by QxMD. Sign in or create an account to discover new knowledge that matter to you.
Remove bar
Read by QxMD icon Read

Search Tips

Use Boolean operators: AND/OR

diabetic AND foot
diabetes OR diabetic

Exclude a word using the 'minus' sign

Virchow -triad

Use Parentheses

water AND (cup OR glass)

Add an asterisk (*) at end of a word to include word stems

Neuro* will search for Neurology, Neuroscientist, Neurological, and so on

Use quotes to search for an exact phrase

"primary prevention of cancer"
(heart or cardiac or cardio*) AND arrest -"American Heart Association"