Change search
Refine search result
1234567 101 - 150 of 1889
CiteExportLink to result list
Permanent link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Rows per page
  • 5
  • 10
  • 20
  • 50
  • 100
  • 250
Sort
  • Standard (Relevance)
  • Author A-Ö
  • Author Ö-A
  • Title A-Ö
  • Title Ö-A
  • Publication type A-Ö
  • Publication type Ö-A
  • Issued (Oldest first)
  • Issued (Newest first)
  • Created (Oldest first)
  • Created (Newest first)
  • Last updated (Oldest first)
  • Last updated (Newest first)
  • Disputation date (earliest first)
  • Disputation date (latest first)
  • Standard (Relevance)
  • Author A-Ö
  • Author Ö-A
  • Title A-Ö
  • Title Ö-A
  • Publication type A-Ö
  • Publication type Ö-A
  • Issued (Oldest first)
  • Issued (Newest first)
  • Created (Oldest first)
  • Created (Newest first)
  • Last updated (Oldest first)
  • Last updated (Newest first)
  • Disputation date (earliest first)
  • Disputation date (latest first)
Select
The maximal number of hits you can export is 250. When you want to export more records please use the Create feeds function.
  • 101.
    Assabie Lake, Yaregal
    Halmstad University, School of Information Science, Computer and Electrical Engineering (IDE), Halmstad Embedded and Intelligent Systems Research (EIS).
    Multifont recognition System for Ethiopic Script2006Licentiate thesis, comprehensive summary (Other academic)
    Abstract [en]

    In this thesis, we present a general framework for multi-font, multi-size and multi-style Ethiopic character recognition system. We propose structural and syntactic techniques for recognition of Ethiopic characters where the graphically comnplex characters are represented by less complex primitive structures and their spatial interrelationships. For each Ethiopic character, the primitive structures and their spatial interrelationships form a unique set of patterns.

    The interrelationships of primitives are represented by a special tree structure which resembles a binary search tree in the sense that it groups child nodes as left and right, and keeps the spatial position of primitives in orderly manner. For a better computational efficiency, the primitive tree is converted into string pattern using in-order traversal, which generates a base of the alphabet that stores possibly occuring string patterns for each character. The recognition of characters is then achieved by matching the generated patterns with each pattern in a stored knowledge base of characters.

    Structural features are extracted using direction field tensor, which is also used for character segmentation. In general, the recognition system does not need size normalization, thinning or other preprocessing procedures. The only parameter that needs to be adjusted during the recognition process is the size of Gaussian window which should be chosen optimally in relation to font sizes. We also constructed an Ethiopic Document Image Database (EDIDB) from real life documents and the recognition system is tested with respect to variations in font type, size, style, document skewness and document type. Experimental results are reported.

  • 102. Astruc, Marine
    et al.
    Malm, Patrik
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Kumar, Rajesh
    Bengtsson, Ewert
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Cluster detection and field-of-view quality rating: Applied to automated Pap-smear analysis2013In: Proc. 2nd International Conference on Pattern Recognition Applications and Methods, SciTePress, 2013, p. 355-364Conference paper (Refereed)
    Abstract [en]

    Automated cervical cancer screening systems require high resolution analysis of a large number of epithelial cells, involving complex algorithms, mainly analysing the shape and texture of cell nuclei. This can be a very time consuming process. An initial selection of relevant fields-of-view in low resolution images could limit the number of fields to be further analysed at a high resolution. In particular, the detection of cell clusters is of interest for nuclei segmentation improvement, and for diagnostic purpose, malignant and endometrial cells being more prone to stick together in clusters than other cells. In this paper, we propose methods aiming at evaluating the quality of fields-of-view in bright-field microscope images of cervical cells. The approach consists in the construction of neighbourhood graphs using the nuclei as the set of vertices. Transformations are then applied on such graphs in order to highlight the main structures in the image. The methods result in the delineation of regions with varying cell density and the identification of cell clusters. Clustering methods are evaluated using a dataset of manually delineated clusters and compared to a related work.

  • 103.
    Auer, Cornelia
    et al.
    Zuse Institute Berlin, Berlin, Germany.
    Nair, Jaya
    IIIT – Bangalore, Electronics City, Hosur Road, Bangalore, India.
    Zobel, Valentin
    Zuse Institue Berlin, Berlin, Germany.
    Hotz, Ingrid
    Zuse Institue Berlin, Berlin, Germany.
    2D Tensor Field Segmentation2011In: Dagstuhl Follow-Ups, E-ISSN 1868-8977, Vol. 2, p. 17-35Article in journal (Refereed)
    Abstract [en]

    We present a topology-based segmentation as means for visualizing 2D symmetric tensor fields. The segmentation uses directional as well as eigenvalue characteristics of the underlying field to delineate cells of similar (or dissimilar) behavior in the tensor field. A special feature of the resulting cells is that their shape expresses the tensor behavior inside the cells and thus also can be considered as a kind of glyph representation. This allows a qualitative comprehension of important structures of the field. The resulting higher-level abstraction of the field provides valuable analysis. The extraction of the integral topological skeleton using both major and minor eigenvector fields serves as a structural pre-segmentation and renders all directional structures in the field. The resulting curvilinear cells are bounded by tensorlines and already delineate regions of equivalent eigenvector behavior. This pre-segmentation is further adaptively refined to achieve a segmentation reflecting regions of similar eigenvalue and eigenvector characteristics. Cell refinement involves both subdivision and merging of cells achieving a predetermined resolution, accuracy and uniformity of the segmentation. The buildingblocks of the approach can be intuitively customized to meet the demands or different applications. Application to tensor fields from numerical stress simulations demonstrates the effectiveness of our method.

  • 104.
    Augustsson, Louise
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction.
    Study and Analysis of Convolutional Neural Networks for Pedestrian Detection in Autonomous Vehicles2018Independent thesis Advanced level (professional degree), 20 credits / 30 HE creditsStudent thesis
    Abstract [en]

    The automotive industry is heading towards more automation. This puts high demands on many systems like Pedestrian Detection Systems. Such systems need to operate in real time with high accuracy and in embedded systems with limited power, memory resources and compute power. This in turn puts high demands on model size and model design. Lately Convolutional Neural Networks (ConvNets) have dominated the field of object detection and therefore it is reasonable to believe that they are suited for pedestrian detection as well. Therefore, this thesis investigates how ConvNets have been used for pedestrian detection and how such solutions can be implemented in embedded systems on FPGAs (Field Programmable Gate Arrays). The conclusions drawn are that ConvNets indeed perform well on pedestrian detection in terms of accuracy but to a cost of large model sizes and heavy computations. This thesis also comes up with a design proposal of a ConvNet for pedestrian detection with the implementation in an embedded system in mind. The proposed network performs well on pedestrian classification and the performance looks promising for detection as well, but further development is required.

  • 105.
    Aviles, Marcos
    et al.
    GMV, Spain.
    Siozios, Kostas
    School of ECE, National Technical University of Athens, Greece.
    Diamantopoulos, Dionysios
    School of ECE, National Technical University of Athens, Greece.
    Nalpantidis, Lazaros
    Production and Management Engineering Dept., Democritus University of Thrace, Greece.
    Kostavelis, Ioannis
    Production and Management Engineering Dept., Democritus University of Thrace, Greece.
    Boukas, Evangelos
    Production and Management Engineering Dept., Democritus University of Thrace, Greece.
    Soudris, Dimitrios
    School of ECE, National Technical University of Athens, Greece.
    Gasteratos, Antonios
    Production and Management Engineering Dept., Democritus University of Thrace, Greece.
    A co-design methodology for implementing computer vision algorithms for rover navigation onto reconfigurable hardware2011In: Proceedings of the FPL2011 Workshop on Computer Vision on Low-Power Reconfigurable Architectures, 2011, p. 9-10Conference paper (Other academic)
    Abstract [en]

    Vision-based robotics applications have been widely studied in the last years. However, up to now solutions that have been proposed were affecting mostly software level. The SPARTAN project focuses in the tight and optimal implementation of computer vision algorithms targeting to rover navigation. For evaluation purposes, these algorithms will be implemented with a co-design methodology onto a Virtex-6 FPGA device.

  • 106.
    Axelsson, Emil
    et al.
    Linköping University, Department of Science and Technology, Media and Information Technology. Linköping University, Faculty of Science & Engineering.
    Costa, Jonathas
    NYU, NY 10003 USA.
    Silva, Claudio
    NYU, NY 10003 USA.
    Emmart, Carter
    Amer Museum Nat Hist, NY 10024 USA.
    Bock, Alexander
    Linköping University, Department of Science and Technology. Linköping University, Faculty of Science & Engineering.
    Ynnerman, Anders
    Linköping University, Department of Science and Technology, Media and Information Technology. Linköping University, Faculty of Science & Engineering. Linköping University, Center for Medical Image Science and Visualization (CMIV).
    Dynamic Scene Graph: Enabling Scaling, Positioning, and Navigation in the Universe2017In: Computer graphics forum (Print), ISSN 0167-7055, E-ISSN 1467-8659, Vol. 36, no 3, p. 459-468Article in journal (Refereed)
    Abstract [en]

    In this work, we address the challenge of seamlessly visualizing astronomical data exhibiting huge scale differences in distance, size, and resolution. One of the difficulties is accurate, fast, and dynamic positioning and navigation to enable scaling over orders of magnitude, far beyond the precision of floating point arithmetic. To this end we propose a method that utilizes a dynamically assigned frame of reference to provide the highest possible numerical precision for all salient objects in a scene graph. This makes it possible to smoothly navigate and interactively render, for example, surface structures on Mars and the Milky Way simultaneously. Our work is based on an analysis of tracking and quantification of the propagation of precision errors through the computer graphics pipeline using interval arithmetic. Furthermore, we identify sources of precision degradation, leading to incorrect object positions in screen-space and z-fighting. Our proposed method operates without near and far planes while maintaining high depth precision through the use of floating point depth buffers. By providing interoperability with order-independent transparency algorithms, direct volume rendering, and stereoscopy, our approach is well suited for scientific visualization. We provide the mathematical background, a thorough description of the method, and a reference implementation.

  • 107.
    Axelsson, Maria
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Centre for Image Analysis. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    An evaluation of scale and noise sensitivity of fibre orientation estimation in volume images2009In: Image Analysis and Processing - ICIAP 2009, Berlin: Springer , 2009, p. 975-984Conference paper (Refereed)
  • 108.
    Axelsson, Maria
    et al.
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Centre for Image Analysis. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Svensson, Stina
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Centre for Image Analysis. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    3D pore structure characterisation of paper2010In: Pattern Analysis and Applications, ISSN 1433-7541, E-ISSN 1433-755X, Vol. 13, no 2, p. 159-172Article in journal (Refereed)
    Abstract [en]

    Pore structure characterisation of paper, using automated image analysis methods, has previously been performed in two-dimensional images. Three dimensional (3D) images have become available and thereby new representations and corresponding measurements are needed for 3D pore structure characterisation. In this article, we present a new pore structure representation, the individual pore-based skeleton, and new quantitative measurements for individual pores in 3D, such as surface area, orientation, anisotropy, and size distributions. We also present measurements for network relations, like tortuosity and connectivity. The data used to illustrate the pore structure representations and corresponding measurements are high resolution X-ray microtomography volume images of a layered duplex board imaged at the European Synchrotron Radiation Facility (ESRF). Quantification of the pore structure is exemplified and the results show that differences in pore structure between the layers in the cardboard can be characterised using the presented methods.

  • 109.
    Axelsson, Maria
    et al.
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Centre for Image Analysis. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Svensson, Stina
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Centre for Image Analysis.
    Borgefors, Gunilla
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Centre for Image Analysis.
    Reduction of Ring Artifacts in High Resolution X-Ray Microtomography Images2006In: Pattern Recognition: 28th DAGM Symposium, Berlin, Germany, September 2006, Proceedings, 2006, p. 61-70Conference paper (Refereed)
    Abstract [en]

    Ring artifacts can occur in reconstructed images from X-ray microtomography as full or partial circles centred on the rotation axis. In this paper, a 2D method is proposed that reduces these ring artifacts in the reconstructed images. The method consists of two main parts. First, the artifacts are localised in the image using local orientation estimation of the image structures and filtering to find ring patterns in the orientation information. Second, the map of the located artifacts is used to calculate a correction image using normalised convolution. The method is evaluated on 2D images from volume data of paper fibre imaged at the European Synchrotron Radiation Facility (ESRF) with high resolution X-ray microtomography. The results show that the proposed method reduces the artifacts and restores the pixel values for all types of partial and complete ring artifacts where the signal is not completely saturated.

  • 110.
    Axelsson, Maria
    et al.
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Centre for Image Analysis. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Östlund, Catherine
    Vomhoff, Hannes
    Svensson, Stina
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Centre for Image Analysis.
    Estimation of the pore volume at the interface between paper web and press felt2006In: Nordic Pulp & Paper Research Journal, ISSN 0283-2631, E-ISSN 2000-0669, Vol. 21, no 3, p. 395-402Article in journal (Refereed)
    Abstract [en]

    A method for determining the water content at the interface between a press felt and a paper web has been developed. The water content was obtained by subtracting the estimated volume of the indented fibre web from the measured felt surface porosity of the press felt. The felt surface porosity was calculated from a topography map that was imaged with a Confocal Laser Scanning Microscope (CLSM) method. Here, the press felt was compressed against a smooth surface using a stress in the range of 0 to 10 MPa. Artefacts in the CLSM images were reduced using an image analysis method. The indentation of paper webs into the measured felt surface pores at different applied pressures was estimated using another image analysis method, simulating a rolling ball, with different radii of curvature for the different pressures and grammages, rolling over the felt surface. The ball radii were determined for a low and a high grammage web using the STFI-Packforsk Dewatering model. The method was evaluated in a case study with four press felts that had batt fibre diameters in a range between 22 and 78 μm. The indentation was calculated for webs with a low (15 g/m2) and a high grammage (105 g/m2), respectively. The evaluation showed that a considerable amount of porespace is available at the interface between the web and the felt. In most cases, the volume of the water-filled pores accounted for approximately 50% of the total surface porosity of the felt. Assuming a complete water saturation of the web/felt interface, approximately 10 g/m2 of water for the finest felt surface up to 40 g/m2 for the coarsest felt surface, could be located at the interface between the press felt and the paper web at a load of 10 MPa. This implies that a considerable amount of water is available for separation rewetting.

  • 111.
    Ayyalasomayajula, Kalyan Ram
    et al.
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Brun, Anders
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Document Binarization Combining with Graph Cuts and Deep Neural Networks2017Conference paper (Other academic)
  • 112.
    Ayyalasomayajula, Kalyan Ram
    et al.
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Brun, Anders
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Historical document binarization combining semantic labeling and graph cuts2017In: Image Analysis: Part I, Springer, 2017, p. 386-396Conference paper (Refereed)
    Abstract [en]

    Most data mining applications on collections of historical documents require binarization of the digitized images as a pre-processing step. Historical documents are often subjected to degradations such as parchment aging, smudges and bleed through from the other side. The text is sometimes printed, but more often handwritten. Mathematical modeling of appearance of the text, background and all kinds of degradations, is challenging. In the current work we try to tackle binarization as pixel classification problem. We first apply semantic segmentation, using fully convolutional neural networks. In order to improve the sharpness of the result, we then apply a graph cut algorithm. The labels from the semantic segmentation are used as approximate estimates of the text and background, with the probability map of background used for pruning the edges in the graph cut. The results obtained show significant improvement over the state of the art approach.

  • 113.
    Ayyalasomayajula, Kalyan Ram
    et al.
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Brun, Anders
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Semantic Labeling using Convolutional Networks coupled with Graph-Cuts for Document binarization2017Conference paper (Other academic)
  • 114.
    Ayyalasomayajula, Kalyan Ram
    et al.
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Brun, Anders
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Topological clustering guided document binarization2015Report (Other academic)
    Abstract [en]

    The current approach for text binarization proposes a clustering algorithm as a preprocessing stage to an energy-based segmentation method. It uses a clustering algorithm to obtain a coarse estimate of the background (BG) and foreground (FG) pixels. These estimates are usedas a prior for the source and sink points of a graph cut implementation, which is used to efficiently find the minimum energy solution of an objective function to separate the BG and FG. The binary image thus obtained is used to refine the edge map that guides the graph cut algorithm. A final binary image is obtained by once again performing the graph cut guided by the refined edges on Laplacian of the image.

  • 115.
    Ayyalasomayajula, Kalyan Ram
    et al.
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Malmberg, Filip
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Brun, Anders
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    PDNet: Semantic segmentation integrated with a primal-dual network for document binarization2019In: Pattern Recognition Letters, ISSN 0167-8655, E-ISSN 1872-7344, Vol. 121, p. 52-60Article in journal (Refereed)
    The full text will be freely available from 2020-05-17 16:13
  • 116.
    Ayyalasomayajula, Kalyan Ram
    et al.
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Nettelblad, Carl
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Scientific Computing.
    Brun, Anders
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.
    Feature evaluation for handwritten character recognition with regressive and generative Hidden Markov Models2016In: Advances in Visual Computing: Part I, Springer, 2016, p. 278-287Conference paper (Refereed)
  • 117.
    Azizpour, Hossein
    et al.
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Laptev, I.
    Object detection using strongly-supervised deformable part models2012In: Computer Vision – ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, October 7-13, 2012, Proceedings, Part I / [ed] Andrew Fitzgibbon, Svetlana Lazebnik, Pietro Perona, Yoichi Sato, Cordelia Schmid, Springer, 2012, no PART 1, p. 836-849Conference paper (Refereed)
    Abstract [en]

    Deformable part-based models [1, 2] achieve state-of-the-art performance for object detection, but rely on heuristic initialization during training due to the optimization of non-convex cost function. This paper investigates limitations of such an initialization and extends earlier methods using additional supervision. We explore strong supervision in terms of annotated object parts and use it to (i) improve model initialization, (ii) optimize model structure, and (iii) handle partial occlusions. Our method is able to deal with sub-optimal and incomplete annotations of object parts and is shown to benefit from semi-supervised learning setups where part-level annotation is provided for a fraction of positive examples only. Experimental results are reported for the detection of six animal classes in PASCAL VOC 2007 and 2010 datasets. We demonstrate significant improvements in detection performance compared to the LSVM [1] and the Poselet [3] object detectors.

  • 118.
    Azizpour, Hossein
    et al.
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Razavian, Ali Sharif
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Sullivan, Josephine
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Maki, Atsuto
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Carlsson, Stefan
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    From Generic to Specific Deep Representations for Visual Recognition2015In: Proceedings of CVPR 2015, IEEE conference proceedings, 2015Conference paper (Refereed)
    Abstract [en]

    Evidence is mounting that ConvNets are the best representation learning method for recognition. In the common scenario, a ConvNet is trained on a large labeled dataset and the feed-forward units activation, at a certain layer of the network, is used as a generic representation of an input image. Recent studies have shown this form of representation to be astoundingly effective for a wide range of recognition tasks. This paper thoroughly investigates the transferability of such representations w.r.t. several factors. It includes parameters for training the network such as its architecture and parameters of feature extraction. We further show that different visual recognition tasks can be categorically ordered based on their distance from the source task. We then show interesting results indicating a clear correlation between the performance of tasks and their distance from the source task conditioned on proposed factors. Furthermore, by optimizing these factors, we achieve stateof-the-art performances on 16 visual recognition tasks.

  • 119.
    Azizpour, Hossein
    et al.
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Sharif Razavian, Ali
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Sullivan, Josephine
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Maki, Atsuto
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Carlssom, Stefan
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Factors of Transferability for a Generic ConvNet Representation2016In: IEEE Transaction on Pattern Analysis and Machine Intelligence, ISSN 0162-8828, E-ISSN 1939-3539, Vol. 38, no 9, p. 1790-1802, article id 7328311Article in journal (Refereed)
    Abstract [en]

    Evidence is mounting that Convolutional Networks (ConvNets) are the most effective representation learning method for visual recognition tasks. In the common scenario, a ConvNet is trained on a large labeled dataset (source) and the feed-forward units activation of the trained network, at a certain layer of the network, is used as a generic representation of an input image for a task with relatively smaller training set (target). Recent studies have shown this form of representation transfer to be suitable for a wide range of target visual recognition tasks. This paper introduces and investigates several factors affecting the transferability of such representations. It includes parameters for training of the source ConvNet such as its architecture, distribution of the training data, etc. and also the parameters of feature extraction such as layer of the trained ConvNet, dimensionality reduction, etc. Then, by optimizing these factors, we show that significant improvements can be achieved on various (17) visual recognition tasks. We further show that these visual recognition tasks can be categorically ordered based on their similarity to the source task such that a correlation between the performance of tasks and their similarity to the source task w.r.t. the proposed factors is observed.

  • 120.
    Bacciu, Davide
    et al.
    Università di Pisa, Pisa, Italy.
    Di Rocco, Maurizio
    Örebro University, Örebro, Sweden.
    Dragone, Mauro
    Heriot-Watt University, Edinburgh, UK.
    Gallicchio, Claudio
    Università di Pisa, Pisa, Italy.
    Micheli, Alessio
    Università di Pisa, Pisa, Italy.
    Saffiotti, Alessandro
    Örebro University, School of Science and Technology.
    An ambient intelligence approach for learning in smart robotic environments2019In: Computational intelligence, ISSN 0824-7935, E-ISSN 1467-8640Article in journal (Refereed)
    Abstract [en]

    Smart robotic environments combine traditional (ambient) sensing devices and mobile robots. This combination extends the type of applications that can be considered, reduces their complexity, and enhances the individual values of the devices involved by enabling new services that cannot be performed by a single device. To reduce the amount of preparation and preprogramming required for their deployment in real-world applications, it is important to make these systems self-adapting. The solution presented in this paper is based upon a type of compositional adaptation where (possibly multiple) plans of actions are created through planning and involve the activation of pre-existing capabilities. All the devices in the smart environment participate in a pervasive learning infrastructure, which is exploited to recognize which plans of actions are most suited to the current situation. The system is evaluated in experiments run in a real domestic environment, showing its ability to proactively and smoothly adapt to subtle changes in the environment and in the habits and preferences of their user(s), in presence of appropriately defined performance measuring functions.

  • 121.
    Baisero, Andrea
    et al.
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Pokorny, Florian T.
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Kragic, Danica
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Ek, Carl Henrik
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    The Path Kernel2013In: ICPRAM 2013 - Proceedings of the 2nd International Conference on Pattern Recognition Applications and Methods, 2013, p. 50-57Conference paper (Refereed)
    Abstract [en]

    Kernel methods have been used very successfully to classify data in various application domains. Traditionally, kernels have been constructed mainly for vectorial data defined on a specific vector space. Much less work has been addressing the development of kernel functions for non-vectorial data. In this paper, we present a new kernel for encoding sequential data. We present our results comparing the proposed kernel to the state of the art, showing a significant improvement in classification and a much improved robustness and interpretability.

  • 122.
    Bajic, Buda
    et al.
    Faculty of Technical Sciences, University of Novi Sad, Serbia.
    Lindblad, Joakim
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction. Mathematical Institute, Serbian Academy of Sciences and Arts, Belgrade, Serbia.
    Sladoje, Natasa
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Mathematical Institute, Serbian Academy of Sciences and Arts, Belgrade, Serbia.
    Single image super-resolution reconstruction in presence of mixed Poisson-Gaussian noise2016In: 2016 SIXTH INTERNATIONAL CONFERENCE ON IMAGE PROCESSING THEORY, TOOLS AND APPLICATIONS (IPTA), IEEE, 2016Conference paper (Refereed)
    Abstract [en]

    Single image super-resolution (SR) reconstructionaims to estimate a noise-free and blur-free high resolution imagefrom a single blurred and noisy lower resolution observation.Most existing SR reconstruction methods assume that noise in theimage is white Gaussian. Noise resulting from photon countingdevices, as commonly used in image acquisition, is, however,better modelled with a mixed Poisson-Gaussian distribution. Inthis study we propose a single image SR reconstruction methodbased on energy minimization for images degraded by mixedPoisson-Gaussian noise.We evaluate performance of the proposedmethod on synthetic images, for different levels of blur andnoise, and compare it with recent methods for non-Gaussiannoise. Analysis shows that the appropriate treatment of signaldependentnoise, provided by our proposed method, leads tosignificant improvement in reconstruction performance.

  • 123.
    Bajic, Buda
    et al.
    Univ Novi Sad, Fac Tech Sci, Novi Sad, Serbia.
    Lindblad, Joakim
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction. Serbian Acad Arts & Sci, Math Inst, Belgrade, Serbia.
    Sladoje, Natasa
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction. Serbian Acad Arts & Sci, Math Inst, Belgrade, Serbia.
    Sparsity promoting super-resolution coverage segmentation by linear unmixing in presence of blur and noise2019In: Journal of Electronic Imaging (JEI), ISSN 1017-9909, E-ISSN 1560-229X, Vol. 28, no 1, article id 013046Article in journal (Refereed)
    Abstract [en]

    We present a segmentation method that estimates the relative coverage of each pixel in a sensed image by each image component. The proposed super-resolution blur-aware model (utilizes a priori knowledge of the image blur) for linear unmixing of image intensities relies on a sparsity promoting approach expressed by two main requirements: (i) minimization of Huberized total variation, providing smooth object boundaries and noise removal, and (ii) minimization of nonedge image fuzziness, responding to an assumption that imaged objects are crisp and that fuzziness is mainly due to the imaging and digitization process. Edge fuzziness due to partial coverage is allowed, enabling subpixel precise feature estimates. The segmentation is formulated as an energy minimization problem and solved by the spectral projected gradient method, utilizing a graduated nonconvexity scheme. Quantitative and qualitative evaluation on synthetic and real multichannel images confirms good performance, particularly relevant when subpixel precision in segmentation and subsequent analysis is a requirement. (C) 2019 SPIE and IS&T

  • 124.
    Bajic, Buda
    et al.
    Faculty of Technical Sciences, University of Novi Sad, Serbia.
    Lindblad, Joakim
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction. Serbian Acad Arts & Sci, Math Inst, Belgrade, Serbia.
    Sladoje, Nataša
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Serbian Acad Arts & Sci, Math Inst, Belgrade, Serbia.
    Blind restoration of images degraded with mixed poisson-Gaussian noise with application in transmission electron microscopy2016In: 2016 Ieee 13Th International Symposium On Biomedical Imaging (ISBI), IEEE, 2016, p. 123-127Conference paper (Refereed)
    Abstract [en]

    Noise and blur, present in images after acquisition, negatively affect their further analysis. For image enhancement when the Point Spread Function (PSF) is unknown, blind deblurring is suitable, where both the PSF and the original image are simultaneously reconstructed. In many realistic imaging conditions, noise is modelled as a mixture of Poisson (signal-dependent) and Gaussian (signal independent) noise. In this paper we propose a blind deconvolution method for images degraded by such mixed noise. The method is based on regularized energy minimization. We evaluate its performance on synthetic images, for different blur kernels and different levels of noise, and compare with non-blind restoration. We illustrate the performance of the method on Transmission Electron Microscopy images of cilia, used in clinical practice for diagnosis of a particular type of genetic disorders.

  • 125.
    Bajic, Buda
    et al.
    Faculty of Technical Sciences, University of Novi Sad, Serbia.
    Lindblad, Joakim
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction. Mathematical Institute, Serbian Academy of Sciences and Arts, Belgrade, Serbia.
    Sladoje, Nataša
    Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Mathematical Institute, Serbian Academy of Sciences and Arts, Belgrade, Serbia.
    Restoration of images degraded by signal-dependent noise based on energy minimization: an empirical study2016In: Journal of Electronic Imaging (JEI), ISSN 1017-9909, E-ISSN 1560-229X, Vol. 25, no 4, article id 043020Article in journal (Refereed)
    Abstract [en]

    Most energy minimization-based restoration methods are developed for signal-independent Gaussian noise. The assumption of Gaussian noise distribution leads to a quadratic data fidelity term, which is appealing in optimization. When an image is acquired with a photon counting device, it contains signal-dependent Poisson or mixed Poisson–Gaussian noise. We quantify the loss in performance that occurs when a restoration method suited for Gaussian noise is utilized for mixed noise. Signal-dependent noise can be treated by methods based on either classical maximum a posteriori (MAP) probability approach or on a variance stabilization approach (VST). We compare performances of these approaches on a large image material and observe that VST-based methods outperform those based on MAP in both quality of restoration and in computational efficiency. We quantify improvement achieved by utilizing Huber regularization instead of classical total variation regularization. The conclusion from our study is a recommendation to utilize a VST-based approach combined with regularization by Huber potential for restoration of images degraded by blur and signal-dependent noise. This combination provides a robust and flexible method with good performance and high speed.

  • 126.
    Baldassarre, Federico
    et al.
    KTH, School of Electrical Engineering and Computer Science (EECS), Robotics, Perception and Learning, RPL.
    Azizpour, Hossein
    KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP.
    Explainability Techniques for Graph Convolutional Networks2019Conference paper (Refereed)
    Abstract [en]

    Graph Networks are used to make decisions in potentially complex scenarios but it is usually not obvious how or why they made them. In this work, we study the explainability of Graph Network decisions using two main classes of techniques, gradient-based and decomposition-based, on a toy dataset and a chemistry task. Our study sets the ground for future development as well as application to real-world problems.

  • 127.
    Ballerini, L.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    A Simple Method to Measure Homogeneity of Fat Distribution in Meat2001Conference paper (Refereed)
    Abstract [en]

    Fat distribution is an important criterium for meat quality evaluation and

  • 128.
    Ballerini, L.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Detection and quantification of foveal avascular zone alterations in diabetic retinopathy2000In: 1st Int. Workshop on Computer Assisted Fundus Image Analysis (CAFIA), 2000Conference paper (Refereed)
    Abstract [en]

    In this work a computational approach for detecting and quantifying diabetic

  • 129.
    Ballerini, L.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Determination of fat content in NMR images of meat2000Conference paper (Refereed)
    Abstract [en]

    In this paper we present an application to food science of image processing

  • 130.
    Ballerini, L.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Determination of fat contents in NMR images of meat: preliminary results2000In: Symposium on Image Analysis - SSAB 2000, 2000, p. 79-82Conference paper (Other scientific)
  • 131.
    Ballerini, L.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Genetic Snakes for Color Image Segmentation2001Conference paper (Refereed)
    Abstract [en]

    The world of meat faces a permanent need for new methods of meat quality

  • 132.
    Ballerini, L.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    How Do People Choose Meat?2001In: Swedish Society for Automated Image Analysis Symposium - SSAB 2001,ITN, Campus Norrköping, LinköpingUniversity, 2001, p. 119-122Conference paper (Other scientific)
  • 133.
    Ballerini, L.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Image Analysis for the Food Industry: Digital Camera Photographs and Nuclear Magnetic Resonance Images2001In: Electronic Imaging, Vol. 11, no 2, p. 7-Article in journal (Refereed)
  • 134.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Barone, L.T.
    Bianchetti, M.
    Monforti Ferrario, F.
    Sacca', F.
    Usai, C.
    Cervelli in fuga (Brains on the run - Stories of Italian researchers fled abroad)2001Book (Other scientific)
  • 135.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Bocchi, L.
    A Fractal Approach to Predict Fat Content in Meat Images2001Conference paper (Refereed)
    Abstract [en]

    Intramuscular fat content in meat influences some important meat quality

  • 136.
    Ballerini. L., Bocchi
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    L.,
    Segmentation of liver images by texture and genetic snakes2002Conference paper (Refereed)
  • 137.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Bocchi, L.
    Hullberg, A.
    Determination of Pores in Pig Meat Images2002In: International Conference on Computer Vision and Graphics, Zakopane, Poland, 2002, p. 70-78Conference paper (Refereed)
    Abstract [en]

    In this paper we present an image processing application for

  • 138.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Borgefors, G.
    Theory and Applications of Image Analysis at the Centre for Image Analysis2001In: 5th Korea-Germany JointWorkshop on Advanced Medical Image Processing, Seoul, Korea, 2001Conference paper (Other scientific)
  • 139.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Hullberg, A.
    Determination of holes in pig meat images2002In: Proceedings SSAB'02 Symposium on Image Analysis, 2002, p. 53-56Conference paper (Other scientific)
    Abstract [en]

    In this paper we present an image processing application for

  • 140.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Högberg, A.
    How Do People Choose Meat?2001Conference paper (Refereed)
    Abstract [en]

    In this paper we present a survey carried out to understand the choice of

  • 141.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Högberg, A.
    Borgefors., G.
    Bylund, A.-C.
    Lindgård, A.
    Lundström, K.
    Rakotonirainy, O.
    Soussi, B.
    A Segmentation Technique to Determine Fat Content in NMR Images of Beef Meat2002In: IEEE Transactions on Nuclear Science, Vol. 49, no 1, p. 195-199Article in journal (Refereed)
    Abstract [en]

    The world of meat faces a permanent need for new methods of meat

  • 142.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Högberg, A.
    Borgefors, G.
    Bylund, A.-C.
    Lindgård, A.
    Lundström, K.
    Rakotonirainy, O.
    Soussi, B.
    Testing MRI and image analysis techniques for fat quantification in meat science2000Conference paper (Refereed)
    Abstract [en]

    The world of meat faces a permanent need for new methods of meat quality

  • 143.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Högberg, A.
    Lundström, K.
    Borgefors, G.
    Colour Image Analysis Technique for Measuring of Fat in Meat: An Application forthe Meat Industry2001Conference paper (Refereed)
    Abstract [en]

    Intramuscular fat content in meat influences some important meat quality

  • 144.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Piazza, E.
    A picture of doctoral studies in Italy2001In: Eurodoc 2001, European Conference of Doctoral Students, Uppsala, Sweden, 2001Conference paper (Other scientific)
  • 145.
    Ballerini, L.
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Piazza, E.
    The future of Italian doctors2002In: Eurodoc 2002, European Conference of Doctoral Students, Girona, Spain, 2002Conference paper (Other scientific)
  • 146. Barekatain, M.
    et al.
    Marti, Miquel
    KTH. Polytechnic University of Catalonia, Spain.
    Shih, H. -F
    Murray, Samuel
    KTH, School of Computer Science and Communication (CSC).
    Nakayama, K.
    Matsuo, Y.
    Prendinger, H.
    Okutama-Action: An Aerial View Video Dataset for Concurrent Human Action Detection2017In: 30th IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2017, IEEE Computer Society, 2017, Vol. 2017, p. 2153-2160Conference paper (Refereed)
    Abstract [en]

    Despite significant progress in the development of human action detection datasets and algorithms, no current dataset is representative of real-world aerial view scenarios. We present Okutama-Action, a new video dataset for aerial view concurrent human action detection. It consists of 43 minute-long fully-annotated sequences with 12 action classes. Okutama-Action features many challenges missing in current datasets, including dynamic transition of actions, significant changes in scale and aspect ratio, abrupt camera movement, as well as multi-labeled actors. As a result, our dataset is more challenging than existing ones, and will help push the field forward to enable real-world applications.

  • 147.
    Barkman, Richard Dan William
    Karlstad University, Faculty of Health, Science and Technology (starting 2013).
    Object Tracking Achieved by Implementing Predictive Methods with Static Object Detectors Trained on the Single Shot Detector Inception V2 Network2019Independent thesis Advanced level (degree of Master (Two Years)), 20 credits / 30 HE creditsStudent thesis
    Abstract [en]

    In this work, the possibility of realising object tracking by implementing predictive methods with static object detectors is explored. The static object detectors are obtained as models trained on a machine learning algorithm, or in other words, a deep neural network. Specifically, it is the single shot detector inception v2 network that will be used to train such models. Predictive methods will be incorporated to the end of improving the obtained models’ precision, i.e. their performance with respect to accuracy. Namely, Lagrangian mechanics will be employed to derived equations of motion for three different scenarios in which the object is to be tracked. These equations of motion will be implemented as predictive methods by discretising and combining them with four different iterative formulae.

    In ch. 1, the fundamentals of supervised machine learning, neural networks, convolutional neural networks as well as the workings of the single shot detector algorithm, approaches to hyperparameter optimisation and other relevant theory is established. This includes derivations of the relevant equations of motion and the iterative formulae with which they were implemented. In ch. 2, the experimental set-up that was utilised during data collection, and the manner by which the acquired data was used to produce training, validation and test datasets is described. This is followed by a description of how the approach of random search was used to train 64 models on 300×300 datasets, and 32 models on 512×512 datasets. Consecutively, these models are evaluated based on their performance with respect to camera-to-object distance and object velocity. In ch. 3, the trained models were verified to possess multi-scale detection capabilities, as is characteristic of models trained on the single shot detector network. While the former is found to be true irrespective of the resolution-setting of the dataset that the model has been trained on, it is found that the performance with respect to varying object velocity is significantly more consistent for the lower resolution models as they operate at a higher detection rate.

    Ch. 3 continues with that the implemented predictive methods are evaluated. This is done by comparing the resulting deviations when they are let to predict the missing data points from a collected detection pattern, with varying sampling percentages. It is found that the best predictive methods are those that make use of the least amount of previous data points. This followed from that the data upon which evaluations were made contained an unreasonable amount of noise, considering that the iterative formulae implemented do not take noise into account. Moreover, the lower resolution models were found to benefit more than those trained on the higher resolution datasets because of the higher detection frequency they can employ.

    In ch. 4, it is argued that the concept of combining predictive methods with static object detectors to the end of obtaining an object tracker is promising. Moreover, the models obtained on the single shot detector network are concluded to be good candidates for such applications. However, the predictive methods studied in this thesis should be replaced with some method that can account for noise, or be extended to be able to account for it. A profound finding is that the single shot detector inception v2 models trained on a low-resolution dataset were found to outperform those trained on a high-resolution dataset in certain regards due to the higher detection rate possible on lower resolution frames. Namely, in performance with respect to object velocity and in that predictive methods performed better on the low-resolution models.

  • 148.
    Barnada, Marc
    et al.
    Linköping University, Department of Electrical Engineering, Computer Vision. Linköping University, Faculty of Science & Engineering. Goethe University of Frankfurt, Germany.
    Conrad, Christian
    Goethe University of Frankfurt, Germany.
    Bradler, Henry
    Goethe University of Frankfurt, Germany.
    Ochs, Matthias
    Goethe University of Frankfurt, Germany.
    Mester, Rudolf
    Linköping University, Department of Electrical Engineering, Computer Vision. Linköping University, Faculty of Science & Engineering. Goethe University of Frankfurt, Germany.
    Estimation of Automotive Pitch, Yaw, and Roll using Enhanced Phase Correlation on Multiple Far-field Windows2015In: 2015 IEEE Intelligent Vehicles Symposium (IV), IEEE , 2015, p. 481-486Conference paper (Refereed)
    Abstract [en]

    The online-estimation of yaw, pitch, and roll of a moving vehicle is an important ingredient for systems which estimate egomotion, and 3D structure of the environment in a moving vehicle from video information. We present an approach to estimate these angular changes from monocular visual data, based on the fact that the motion of far distant points is not dependent on translation, but only on the current rotation of the camera. The presented approach does not require features (corners, edges,...) to be extracted. It allows to estimate in parallel also the illumination changes from frame to frame, and thus allows to largely stabilize the estimation of image correspondences and motion vectors, which are most often central entities needed for computating scene structure, distances, etc. The method is significantly less complex and much faster than a full egomotion computation from features, such as PTAM [6], but it can be used for providing motion priors and reduce search spaces for more complex methods which perform a complete analysis of egomotion and dynamic 3D structure of the scene in which a vehicle moves.

  • 149.
    Barnden, L
    et al.
    Uppsala University, Interfaculty Units, Centre for Image Analysis. Teknisk-naturvetenskapliga vetenskapsområdet, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis.
    Kwiatek, R
    Lau, Y
    Hutton, B
    Thurfjell, L
    Pile, K
    Rowe, C
    Validation of fully automatic brain SPET to MR co-registration2000In: EUROPEAN JOURNAL OF NUCLEAR MEDICINE, ISSN 0340-6997, Vol. 27, no 2, p. 147-154Article in journal (Refereed)
    Abstract [en]

    Fully automatic co-registration of functional to anatomical brain images using information intrinsic to the scans has been validated in a clinical setting for positron emission tomography (PET), but not for single-photon emission tomography (SPET). In thi

  • 150. Baroffio, L.
    et al.
    Cesana, M.
    Redondi, A.
    Tagliasacchi, M.
    Ascenso, J.
    Monteiro, P.
    Eriksson, Emil
    KTH, School of Electrical Engineering (EES), Communication Networks.
    Dan, G.
    Fodor, Viktoria
    KTH, School of Electrical Engineering (EES), Communication Networks.
    GreenEyes: Networked energy-aware visual analysis2015In: 2015 IEEE International Conference on Multimedia and Expo Workshops, ICMEW 2015, IEEE conference proceedings, 2015Conference paper (Refereed)
    Abstract [en]

    The GreenEyes project aims at developing a comprehensive set of new methodologies, practical algorithms and protocols, to empower wireless sensor networks with vision capabilities. The key tenet of this research is that most visual analysis tasks can be carried out based on a succinct representation of the image, which entails both global and local features, while it disregards the underlying pixel-level representation. Specifically, GreenEyes will pursue the following goals: i) energy-constrained extraction of visual features; ii) rate-efficiency modelling and coding of visual feature; iii) networking streams of visual features. This will have a significant impact on several scenarios including, e.g., smart cities and environmental monitoring.

1234567 101 - 150 of 1889
CiteExportLink to result list
Permanent link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf