Ändra sökning
Avgränsa sökresultatet
1234567 151 - 200 av 534
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Träffar per sida
  • 5
  • 10
  • 20
  • 50
  • 100
  • 250
Sortering
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
  • Standard (Relevans)
  • Författare A-Ö
  • Författare Ö-A
  • Titel A-Ö
  • Titel Ö-A
  • Publikationstyp A-Ö
  • Publikationstyp Ö-A
  • Äldst först
  • Nyast först
  • Skapad (Äldst först)
  • Skapad (Nyast först)
  • Senast uppdaterad (Äldst först)
  • Senast uppdaterad (Nyast först)
Markera
Maxantalet träffar du kan exportera från sökgränssnittet är 250. Vid större uttag använd dig av utsökningar.
  • 151.
    Creignou, Nadia
    et al.
    Aix Marseille University, France.
    Egly, Uwe
    Vienna University of Technology, Austria.
    Schmidt, Johannes
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Complexity Classifications for Logic-Based Argumentation2014Ingår i: ACM Transactions on Computational Logic, ISSN 1529-3785, Vol. 15, nr 3, 19- s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    We consider logic-based argumentation in which an argument is a pair (Phi, alpha), where the support Phi is a minimal consistent set of formulae taken from a given knowledge base (usually denoted by Delta) that entails the claim alpha (a formula). We study the complexity of three central problems in argumentation: the existence of a support Phi subset of Delta, the verification of a support, and the relevance problem (given psi, is there a support Phi such that psi is an element of Phi?). When arguments are given in the frill language of propositional logic, these problems are computationally costly tasks: the verification problem is DP-complete; the others are Sigma(P)(2)-complete. We study these problems in Schaefers famous framework where the considered propositional formulae are in generalized conjunctive normal form. This means that formulae are conjunctions of constraints built upon a fixed finite set of Boolean relations Gamma (the constraint language). We show that according to the properties of this language Gamma, deciding whether there exists a support for a claim in a given knowledge base is either polynomial, NP-complete, coNP-complete, or Sigma(P)(2)-complete. We present a dichotomous classification, P or DP-complete, for the verification problem and a trichotomous classification for the relevance problem into either polynomial, NP-complete, or Sigma(P)(2)-complete. These last two classifications are obtained by means of algebraic tools.

  • 152.
    Creignou, Nadia
    et al.
    Aix-Marseille Université, France.
    Meier, Arne
    Leibniz Universität, Hannover, Germany.
    Müller, Julian-Steffen
    Leibniz Universität, Hannover, Germany.
    Schmidt, Johannes
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Vollmer, Heribert
    Leibniz Universität, Hannover, Germany.
    Paradigms for Parameterized Enumeration2013Ingår i: Mathematical Foundations of Computer Science 2013 / [ed] Krishnendu Chatterjee, Jirí Sgall, Springer Berlin/Heidelberg, 2013, 290-301 s.Konferensbidrag (Refereegranskat)
    Abstract [en]

    The aim of the paper is to examine the computational complexity and algorithmics of enumeration, the task to output all solutions of a given problem, from the point of view of parameterized complexity. First we define formally different notions of efficient enumeration in the context of parameterized complexity. Second we show how different algorithmic paradigms can be used in order to get parameter-efficient enumeration algorithms in a number of examples. These paradigms use well-known principles from the design of parameterized decision as well as enumeration techniques, like for instance kernelization and self-reducibility. The concept of kernelization, in particular, leads to a characterization of fixed-parameter tractable enumeration problems.

  • 153.
    Crusoe, Jonathan
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Metoder för användardriven gränssnittsprogrammering2014Självständigt arbete på grundnivå (kandidatexamen), 10 poäng / 15 hpStudentuppsats (Examensarbete)
    Abstract [sv]

    När användare bestämmer sig för att utveckla gränssnitt till sina system sker detta via någon form av verktyg. Vi måste avgöra vilken utvecklingsmetodik som ska användas och hur vi kan tillföra mer funktionalitet för att systemet inte ska bli föråldrat. För att svara på detta bryter vi upp arbetet i två delar. I första delen undersöker vi vilken programmeringsmetodik som lämpar sig bäst för gränssnittsutveckling genom en undersökning i två delar. I andra delen ser vi över vilka lösningar som existerar för att implementera ny funktionalitet till ett verktyg för att sedan presentera en egen lösning.

  • 154.
    Cucurull, Jordi
    et al.
    Linköpings universitet, Tekniska högskolan. Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Asplund, Mikael
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Nadjm-Tehrani, Simin
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Surviving Attacks in Challenged Networks2012Ingår i: IEEE Transactions on Dependable and Secure Computing, ISSN 1545-5971, E-ISSN 1941-0018, Vol. 9, nr 6, 917-929 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    In the event of a disaster, telecommunication infrastructures can be severely damaged or overloaded. Hastily formed networks can provide communication services in an ad hoc manner. These networks are challenging due to the chaotic context where intermittent connection is the norm and the identity and number of participants cannot be assumed. In such environments malicious actors may try to disrupt the communications to create more chaos for their own benefit. This paper proposes a general security framework for monitoring and reacting to disruptive attacks. It includes a collection of functions to detect anomalies, diagnose them, and perform mitigation. The measures are deployed in each node in a fully distributed fashion, but their collective impact is a significant resilience to attacks, so the actors can disseminate information under adverse conditions. The approach is evaluated in the context of a simulated disaster area network with a many-cast dissemination protocol, Random Walk Gossip, with a store-and-forward mechanism. A challenging threat model where adversaries may 1) try to drain the resources both at node level (battery life) and network level (bandwidth), or 2) reduce message dissemination in their vicinity, without spending much of their own energy, is adopted. The results demonstrate that the approach diminishes the impact of the attacks considerably.

  • 155.
    Cucurull, Jordi
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Nadjm-Tehrani, Simin
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Raciti, Massimiliano
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Modular Anomaly Detection for Smartphone Ad Hoc Communication2012Ingår i: Information Security Technology for Applications: 16th Nordic Conference on Secure IT Systems, NordSec 2011, Tallinn, Estonia, October 26-28, 2011, Revised Selected Papers / [ed] Peeter Laud, Springer Berlin/Heidelberg, 2012, Vol. 7161, 65-81 s.Kapitel i bok, del av antologi (Refereegranskat)
    Abstract [en]

    The capabilities of the modern smartphones make them the obvious platform for novel mobile applications. The open architectures, however, also create new vulnerabilities. Measures for prevention, detection, and reaction need to be explored with the peculiarities that resource-constrained devices impose. Smartphones, in addition to cellular broadband network capabilities, include WiFi interfaces that can even be deployed to set up a mobile ad hoc network (MANET). While intrusion detection in MANETs is typically evaluated with network simulators, we argue that it is important to implement and test the solutions in real devices to evaluate their resource footprint. This paper presents a modular implementation of an anomaly detection and mitigation mechanism on top of a dissemination protocol for intermittently-connected MANETs. The overhead of the security solution is evaluated in a small testbed based on three Android-based handsets and a laptop. The study shows the feasibility of the statistics-based anomaly detection regime, having low CPU usage, little added latency, and acceptable memory footprint.

  • 156.
    Dag, Antymos
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Autonomous Indoor Navigation System for Mobile Robots2016Självständigt arbete på grundnivå (kandidatexamen), 10,5 poäng / 16 hpStudentuppsats (Examensarbete)
    Abstract [en]

    With an increasing need for greater traffic safety, there is an increasing demand for means by which solutions to the traffic safety problem can be studied. The purpose of this thesis is to investigate the feasibility of using an autonomous indoor navigation system as a component in a demonstration system for studying cooperative vehicular scenarios. Our method involves developing and evaluating such a navigation system. Our navigation system uses a pre-existing localization system based on passive RFID, odometry and a particle filter. The localization system is used to estimate the robot pose, which is used to calculate a trajectory to the goal. A control system with a feedback loop is used to control the robot actuators and to drive the robot to the goal.

     

    The results of our evaluation tests show that the system generally fulfills the performance requirements stated for the tests. There is however some uncertainty about the consistency of its performance. Results did not indicate that this was caused by the choice of localization techniques. The conclusion is that an autonomous navigation system using the aforementioned localization techniques is plausible for use in a demonstration system. However, we suggest that the system is further tested and evaluated before it is used with applications where accuracy is prioritized.

  • 157.
    Danylenko, Antonina
    et al.
    Linnaeus University, Växjö.
    Löwe, Welf
    Linnaeus University, Växjö.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Comparing Machine Learning Approaches for Context-Aware Composition2011Ingår i: Software Composition / [ed] Sven Apel, Ethan Jackson, Springer, 2011, 18-33 s.Konferensbidrag (Refereegranskat)
    Abstract [en]

    Context-Aware Composition allows to automatically select optimal variants of algorithms, data-structures, and schedules at runtime using generalized dynamic Dispatch Tables. These tables grow exponentially with the number of significant context attributes. To make Context-Aware Composition scale, we suggest four alternative implementations to Dispatch Tables, all well-known in the field of machine learning: Decision Trees, Decision Diagrams, Naive Bayes and Support Vector Machines classifiers. We assess their decision overhead and memory consumption theoretically and practically in a number of experiments on different hardware platforms. Decision Diagrams turn out to be more compact compared to Dispatch Tables, almost as accurate, and faster in decision making. Using Decision Diagrams in Context-Aware Composition leads to a better scalability, i.e., Context-Aware Composition can be applied at more program points and regard more context attributes than before.

  • 158.
    DASH, ASSMITRA
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Minimizing Test Time through Test FlowOptimization in 3D-SICs2013Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    3D stacked ICs (3D-SICs) with multiple dies interconnected by through-silicon-vias(TSVs) are considered as a technology driver and proven to have overwhelming advantagesover traditional ICs with a single die in a package in terms of performance, powerconsumption and silicon overhead. However, these “super chips” bring new challengesto the process of IC manufacturing; among which, testing 3D-SICs is the major andmost complex issue to deal with. In traditional ICs, tests can usually be performedat two stages (test instances), namely: a wafer sort and a package test. Whereas for3D-SICs, tests can be performed after each stacking event where a new die is stackedover a partial stack. This expands the set of available test instances. A combination ofselected test instances where a test is performed (active test instance) is known as a testflow. Test time is a major contributor to the total test cost. Test time changes with theselected test flow. Therefore, choosing a cost effective test flow which will minimizesthe test time is absolutely essential.This thesis focuses on finding an optimal test flow which minimizes the test timefor a given 3D-SIC. A mathematical model has been developed to evaluate the test timeof any test flow. Then a heuristic has been proposed for finding a near optimal test flowwhich minimizes the test time. The performance of this approach in terms of computationtime and efficiency has been compared against the minimum test time obtainedby exhaustive search. The heuristic gives good results compared to exhaustive searchwith much lesser computation time.

  • 159.
    Dastgeer, Usman
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Performance-aware Component Composition for GPU-based systems2014Doktorsavhandling, monografi (Övrigt vetenskapligt)
    Abstract [en]

    This thesis addresses issues associated with efficiently programming modern heterogeneous GPU-based systems, containing multicore CPUs and one or more programmable Graphics Processing Units (GPUs). We use ideas from component-based programming to address programming, performance and portability issues of these heterogeneous systems. Specifically, we present three approaches that all use the idea of having multiple implementations for each computation; performance is achieved/retained either a) by selecting a suitable implementation for each computation on a given platform or b) by dividing the computation work across different implementations running on CPU and GPU devices in parallel.

    In the first approach, we work on a skeleton programming library (SkePU) that provides high-level abstraction while making intelligent  implementation selection decisions underneath either before or during the actual program execution. In the second approach, we develop a composition tool that parses extra information (metadata) from XML files, makes certain decisions online, and, in the end, generates code for making the final decisions at runtime. The third approach is a framework that uses source-code annotations and program analysis to generate code for the runtime library to make the selection decision at runtime. With a generic performance modeling API alongside program analysis capabilities, it supports online tuning as well as complex program transformations.

    These approaches differ in terms of genericity, intrusiveness, capabilities and knowledge about the program source-code; however, they all demonstrate usefulness of component programming techniques for programming GPU-based systems. With experimental evaluation, we demonstrate how all three approaches, although different in their own way, provide good performance on different GPU-based systems for a variety of applications.

  • 160.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap, PELAB - Laboratoriet för programmeringsomgivningar. Linköpings universitet, Tekniska högskolan.
    Enmyren, Johan
    Linköpings universitet, Institutionen för datavetenskap, PELAB - Laboratoriet för programmeringsomgivningar. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Auto-tuning SkePU: A multi-backend skeleton programming framework for multi-GPU systems2011Ingår i: IWMSE '11 Proceedings of the 4th International Workshop on Multicore Software Engineering, New York, NY, USA: Association for Computing Machinery (ACM), 2011, 25-32 s.Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    SkePU is a C++ template library that provides a simple and unified interface for specifying data-parallel computations with the help of skeletons on GPUs using CUDA and OpenCL. The interface is also general enough to support other architectures, and SkePU implements both a sequential CPU and a parallel OpenMP backend. It also supports multi-GPU systems. Currently available skeletons in SkePU include map, reduce, mapreduce, map-with-overlap, maparray, and scan. The performance of SkePU generated code is comparable to that of hand-written code, even for more complex applications such as ODE solving.

    In this paper, we discuss initial results from auto-tuning SkePU using an off-line, machine learning approach where we adapt skeletons to a given platform using training data. The prediction mechanism at execution time uses off-line pre-calculated estimates to construct an execution plan for any desired configuration with minimal overhead. The prediction mechanism accurately predicts execution time for repetitive executions and includes a mechanism to predict execution time for user functions of different complexity. The tuning framework covers selection between different backends as well as choosing optimal parameter values for the selected backend. We will discuss our approach and initial results obtained for different skeletons (map, mapreduce, reduce).

  • 161.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    A Framework for Performance-aware Composition of Applications for GPU-based Systems2013Konferensbidrag (Refereegranskat)
    Abstract [en]

    User-level components of applications can be made performance-aware by annotating them with performance model and other metadata. We present a component model and a composition framework for the performance-aware composition of applications for modern GPU-based systems from such components, which may expose multiple implementation variants. The framework targets the composition problem in an integrated manner, with particular focus on global performance-aware composition across multiple invocations. We demonstrate several key features of our framework relating to performance-aware composition including implementation selection, both with performance characteristics being known (or learned) beforehand as well as cases when they are learned at runtime. We also demonstrate hybrid execution capabilities of our framework on real applications. Furthermore, as an important step towards global composition, we present a bulk composition technique that can make better composition decisions by considering information about upcoming calls along with data flow information extracted from the source program by static analysis, thus improving over the traditional greedy performance-aware policy that only considers the current call for optimization.

  • 162.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    A performance-portable generic component for 2D convolution computations on GPU-based systems2011Ingår i: Fourth Swedish Workshop on Multi-Core Computing MCC-2011: November 23-25, 2011, Linköping University, Linköping, Sweden / [ed] Christoph Kessler, Linköping: Linköping University , 2011, Vol. S. 39-44, 39-44 s.Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    In this paper, we describe our work on providing a generic yet optimized GPU (CUDA/OpenCL) implementation for the 2D MapOverlap skeleton. We explain our implementation with the help  of a 2D convolutilution application, implemented using the newly deveioped skeleton.

  • 163.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    A performance-portable generic component for 2D convolution computations on GPU-based systems2012Ingår i: Proceedings of the Fifth International Workshop on Programmability Issues for Heterogeneous Multicores (MULTIPROG-2012) at the HiPEAC-2012 conference, Paris, Jan. 2012 / [ed] E. Ayguade, B. Gaster, L. Howes, P. Stenström, O. Unsal, 2012Konferensbidrag (Refereegranskat)
    Abstract [en]

    In this paper, we describe our work on providing a generic yet optimized GPU (CUDA/OpenCL) implementation for the 2D MapOverlap skeleton. We explain our implementation with the help of a 2D convolution application, implemented using the newly developed skeleton. The memory (constant and shared memory) and adaptive tiling optimizations are applied and their performance implications are evaluated on different classes of GPUs. We present two different metrics to calculate the optimal tiling factor dynamically in an automated way which helps in retaining best performance without manual tuning while moving to newGPU architectures. With our approach, we can achieve average speedups by a factor of 3.6, 2.3, and 2.4 over an otherwise optimized (without tiling) implementation on NVIDIA C2050, GTX280 and 8800 GT GPUs respectively. Above all, the performance portability is achieved without requiring any manual changes in the skeleton program or the skeleton implementation.

  • 164.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Conditional component composition for GPU-based systems2014Ingår i: Proc. Seventh Workshop on Programmability Issues for Multi-Core Computers (MULTIPROG-2014) at HiPEAC-2014, Vienna, Austria, Jan. 2014, Vienna, Austria: HiPEAC NoE , 2014Konferensbidrag (Refereegranskat)
    Abstract [en]

    User-level components can expose multiple functionally equivalent implementations with different resource requirements and performance characteristics. A composition framework can then choose a suitable implementation for each component invocation guided by an objective function (execution time, energy etc.). In this paper, we describe the idea of conditional composition which enables the component writer to specify constraints on the selectability of a given component implementation based on information about the target system and component call properties. By incorporating such information, more informed and user-guided composition decisions can be made and thus more efficient code be generated, as shown with an example scenario for a GPU-based system.

  • 165.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Performance-aware Composition Framework for GPU-based Systems2015Ingår i: Journal of Supercomputing, ISSN 0920-8542, E-ISSN 1573-0484, Vol. 71, nr 12, 4646-4662 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    User-level components of applications can be made performance-aware by annotating them with performance model and other metadata. We present a component model and a composition framework for the automatically optimized composition of applications for modern GPU-based systems from such components, which may expose multiple implementation variants. The framework targets the composition problem in an integrated manner, with the ability to do global performance-aware composition across multiple invocations. We demonstrate several key features of our framework relating to performance-aware composition including implementation selection, both with performance characteristics being known (or learned) beforehand as well as cases when they are learned at runtime. We also demonstrate hybrid execution capabilities of our framework on real applications. Furthermore, we present a bulk composition technique that can make better composition decisions by considering information about upcoming calls along with data flow information extracted from the source program by static analysis. The bulk composition improves over the traditional greedy performance aware policy that only considers the current call for optimization.

  • 166.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap. Linköpings universitet, Tekniska fakulteten.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    Smart Containers and Skeleton Programming for GPU-Based Systems2016Ingår i: International journal of parallel programming, ISSN 0885-7458, E-ISSN 1573-7640, Vol. 44, nr 3, 506-530 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    In this paper, we discuss the role, design and implementation of smart containers in the SkePU skeleton library for GPU-based systems. These containers provide an interface similar to C++ STL containers but internally perform runtime optimization of data transfers and runtime memory management for their operand data on the different memory units. We discuss how these containers can help in achieving asynchronous execution for skeleton calls while providing implicit synchronization capabilities in a data consistent manner. Furthermore, we discuss the limitations of the original, already optimizing memory management mechanism implemented in SkePU containers, and propose and implement a new mechanism that provides stronger data consistency and improves performance by reducing communication and memory allocations. With several applications, we show that our new mechanism can achieve significantly (up to 33.4 times) better performance than the initial mechanism for page-locked memory on a multi-GPU based system.

  • 167.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Li, Lu
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Adaptive Implementation Selection in the SkePU Skeleton Programming Library2013Ingår i: Advanced Parallel Processing Technologies (APPT-2013), Proceedings / [ed] Chengyung Wu and Albert Cohen (eds.), 2013, 170-183 s.Konferensbidrag (Refereegranskat)
    Abstract [en]

    In earlier work, we have developed the SkePU skeleton programming library for modern multicore systems equipped with one or more programmable GPUs. The library internally provides four types of implementations (implementation variants) for each skeleton: serial C++, OpenMP, CUDA and OpenCL targeting either CPU or GPU execution respectively. Deciding which implementation would run faster for a given skeleton call depends upon the computation, problem size(s), system architecture and data locality.

    In this paper, we present our work on automatic selection between these implementation variants by an offline machine learning method which generates a compact decision tree with low training overhead. The proposed selection mechanism is flexible yet high-level allowing a skeleton programmer to control different training choices at a higher abstraction level. We have evaluated our optimization strategy with 9 applications/kernels ported to our skeleton library and achieve on average more than 94% (90%) accuracy with just 0.53% (0.58%) training space exploration on two systems. Moreover, we discuss one application scenario where local optimization considering a single skeleton call can prove sub-optimal, and propose a heuristic for bulk implementation selection considering more than one skeleton call to address such application scenarios.

  • 168.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Li, Lu
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    The PEPPHER composition tool: performance-aware composition for GPU-based systems2014Ingår i: Computing, ISSN 0010-485X, E-ISSN 1436-5057, Vol. 96, nr 12, 1195-1211 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    The PEPPHER (EU FP7 project) component model defines the notion of component, interface and meta-data for homogeneous and heterogeneous parallel systems. In this paper, we describe and evaluate the PEPPHER composition tool, which explores the application’s components and their implementation variants, generates the necessary low-level code that interacts with the runtime system, and coordinates the native compilation and linking of the various code units to compose the overall application code to optimize performance. We discuss the concept of smart containers and its benefits for reducing dispatch overhead, exploiting implicit parallelism across component invocations and runtime optimization of data transfers. In an experimental evaluation with several applications, we demonstrate that the composition tool provides a high-level programming front-end while effectively utilizing the task-based PEPPHER runtime system (StarPU) underneath for different usage scenarios on GPU-based systems.

  • 169.
    Dastgeer, Usman
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Li, Lu
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    The PEPPHER Composition Tool: Performance-Aware Dynamic Composition of Applications for GPU-Based Systems2012Ingår i: High Performance Computing, Networking, Storage and Analysis (SCC), 2012 SC Companion, IEEE, 2012, 711-720 s.Konferensbidrag (Refereegranskat)
    Abstract [en]

    The PEPPHER component model defines an environment for annotation of native C/C++ based components for homogeneous and heterogeneous multicore and manycore systems, including GPU and multi-GPU based systems. For the same computational functionality, captured as a component, different sequential and explicitly parallel implementation variants using various types of execution units might be provided, together with metadata such as explicitly exposed tunable parameters. The goal is to compose an application from its components and variants such that, depending on the run-time context, the most suitable implementation variant will be chosen automatically for each invocation. We describe and evaluate the PEPPHER composition tool, which explores the application's components and their implementation variants, generates the necessary low-level code that interacts with the runtime system, and coordinates the native compilation and linking of the various code units to compose the overall application code. With several applications, we demonstrate how the composition tool provides a high-level programming front-end while effectively utilizing the task-based PEPPHER runtime system (StarPU) underneath.

  • 170.
    Delosières, Laurent
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Nadjm-Tehrani, Simin
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    BATMAN Store-and-Forward: the Best of the Two Worlds2012Ingår i: Pervasive Computing and Communications Workshops (PERCOM Workshops), IEEE , 2012, 721-727 s.Konferensbidrag (Refereegranskat)
    Abstract [en]

    The need for communication is highest in disaster scenarios when the infrastructure is also adversely affected. A recent protocol for ad hoc communication, the BATMAN protocol, is dependent on minimal infrastructure, in the form of mesh nodes that are used as access points, or nodes acting as an intermediary in a multi-hop connection. While BATMAN works well in a scenario in which there is a multihop path from senders to receivers at all times, it will drop the packets in intermittently-connected networks. Moreover, although implementation on a device is essential as a proof of concept, performing large scale evaluations requires a simulation platform in which variations in the operating environment can be studied. This paper is about adding the store-and-forward mechanism to the routing component in BATMAN nodes, to overcome intermittent connectivity through mobility. We describe an extension of the protocol, SF-BATMAN, that has been implemented in an interoperable manner with BATMAN, i.e. with no added signaling, and no change of basic BATMAN settings. We have implemented SF-BATMAN in a packet level simulator (NS3), and demonstrated its performance in a scenario that consists of two regions of connectivity: a wellconnected mesh network and a set of sparser subnetworks. We show that the added capability enhances the performance of BATMAN, through an increase of the delivery ratio by 20% with a lower overhead, while it exhibits a similar latency in comparable network scenarios.

  • 171.
    Delshad, Payman
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Behavior Driven Development in a Large-Scale Application: Evaluation of Usage for Developing IFS Applications2016Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    Nowadays, Agile software development methods are often used in large multisite organizations that develop large-scale applications. Behavior Driven Development (BDD) is a relatively new Agile software development process where the development process starts with acceptance tests written in a natural language. The premise of BDD is to create a common and effective process of communication between different roles in a software project to ensure that every activity can be mapped to the business goal of the application. This thesis work aims to find an effective and efficient BDD process and to evaluate its usage in a large-scale application in a large multisite organization through a series of interviews, a controlled experiment, and an online survey. Furthermore, by means of the aforementioned experiment, the study measures the impact of an experimental usage of BDD on testing quality. To discover an effective and efficient BDD process, two alternatives with automated tests that run on different architectural layers, namely client layer and web service layer, were examined. Based on the defined metrics, the alternative with automated tests that ran directly on the web service layer was chosen as the more efficient process which was compared against the existing Agile-based baseline that used automated client tests. The results show that an efficient BDD process improves the testing quality significantly which can, in turn, result in a better overall software quality.

  • 172.
    Delzanno, Giorgio
    et al.
    Università di Genova.
    Rezine, Ahmed
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    A lightweight regular model checking approach for parameterized systems2012Ingår i: International Journal on Software Tools for Technology Transfer (STTT), ISSN 1433-2779, Vol. 14, nr 2, 207-222 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    In recent years, we have designed a lightweight approach to regular model checking specifically designed for parameterized systems with global conditions. Our approach combines the strength of regular languages, used for representing infinite sets of configurations, with symbolic model checking and approximations. In this paper, we give a uniform presentation of several variations of a symbolic backward reachability scheme in which different classes of regular expressions are used in place of BDDs. The classification of the proposed methods is based on the precision of the resulting approximated analysis.

  • 173.
    Disqah, Arash
    et al.
    Faculty of Engineering and Environment, Northumbria University, Newcastle, UK.
    Maheri, Alireza
    Faculty of Engineering and Environment, Northumbria University, Newcastle, UK.
    Busawon, Krishna
    Faculty of Engineering and Environment, Northumbria University, Newcastle, UK.
    Fritzson, Peter
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Standalone DC Microgrids as Complementarity Dynamical Systems: Modeling and Applications.2014Ingår i: Control Engineering Practice, ISSN 0967-0661, Vol. 35, nr 10, 102-112 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    It is well known that, due to bimodal operation as well as existent discontinuous differential states of batteries, standalone microgrids belong to the class of hybrid dynamical systems of non-Filippov type. In this work, however, standalone microgrids are presented as complementarity systems (CSs) of the Filippov type which is then used to develop a multivariable nonlinear model predictive control (NMPC)-based load tracking strategy as well as Modelica models for long-term simulation purposes. The developed load tracker strategy is a multi-source maximum power point tracker (MPPT) that also regulates the DC bus voltage at its nominal value with the maximum of ±2.0% error despite substantial demand and supply variations.

  • 174.
    Drabent, Wlodzimierz
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    A simple correctness proof for magic transformation2012Ingår i: Theory and Practice of Logic Programming, ISSN 1471-0684, E-ISSN 1475-3081, Vol. 12, nr 6, 929-936 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    The paper presents a simple and concise proof of correctness of the magic transformation. We believe that it may provide a useful example of formal reasoning about logic programs. The correctness property concerns the declarative semantics. The proof, however, refers to the operational semantics (LD-resolution) of the source programs. Its conciseness is due to applying a suitable proof method.

  • 175.
    Drabent, Wlodzimierz
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan. Institute of Computer Science, Polish Academy of Sciences.
    Logic + Control: An Example2012Ingår i: Technical Communications of the 28th International Conference on Logic Programming (ICLP'12) / [ed] Agostino Dovier and Vítor Santos Costa, Dagstuhl Publishing , 2012, 301-311 s.Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    We present a Prolog program (the SAT solver of Howe and King) as a logic program with added control. The control consists of a selection rule (delays of Prolog) and pruning the search space. We construct the logic program together with proofs of its correctness and completeness, with respect to a formal specication. This is augmented by a proof of termination under any selection rule. Correctness and termination are inherited by the Prolog program, the change of selection rule preserves completeness. We prove that completeness is also preserved by one case of pruning; for the other an informal justication is presented.

    For proving correctness we use a method, which should be well known but is often neglected. A contribution of this paper is a method for proving completeness. In particular we introduce a notion of semi-completeness, for which a local sucient condition exists.

    We compare the proof methods with declarative diagnosis (algorithmic debugging). We introduce a method of proving that a certain kind of pruning preserves completeness. We argue that the proof methods correspond to natural declarative thinking about programs, and that they can be used, formally or informally, in every-day programming.

  • 176.
    Drabent, Wlodzimierz
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    Logic + control: On program construction and verification2017Ingår i: Theory and Practice of Logic Programming, ISSN ISSN: 1471-0684 (Print), 1475-3081 (Online)Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    This paper presents an example of formal reasoning about the semantics of a Prolog program of practical importance (the SAT solver of Howe and King). The program is treated as a definite clause logic program with added control. The logic program is constructed by means of stepwise refinement, hand in hand with its correctness and completeness proofs. The proofs are declarative – they do not refer to any operational semantics. Each step of the logic program construction follows a systematic approach to constructing programs which are provably correct and complete. We also prove that correctness and completeness of the logic program is preserved in the final Prolog program. Additionally, we prove termination, occur-check freedom and non-floundering.

    Our example shows how dealing with “logic” and with “control” can be separated. Most of the proofs can be done at the “logic” level, abstracting from any operational semantics.

    The example employs approximate specifications; they are crucial in simplifying reasoning about logic programs. It also shows that the paradigm of semantics-preserving program transformations may be not sufficient.  We suggest considering transformations which preserve correctness and completeness with respect to an approximate specification.

  • 177.
    Drabent, Wlodzimierz
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    On completeness of logic programs2015Ingår i: Logic-Based Program Synthesis and Transformation - 24th International Symposium, (LOPSTR-2014), 2015, Vol. 8981, 261-278 s.Konferensbidrag (Refereegranskat)
    Abstract [en]

    Program correctness (in imperative and functional programming) splits in logic programming into correctness and completeness. Completeness means that a program produces all the answers required by its specification. Little work has been devoted to reasoning about completeness. This paper presents a few sufficient conditions for completeness of definite programs. We also study preserving completeness under some cases of pruning of SLD-trees (e.g. due to using the cut). We treat logic programming as a declarative paradigm, abstracting from any operational semantics as far as possible. We argue that the proposed methods are simple enough to be applied, possibly at an informal level, in practical Prolog programming. We point out importance of approximate specifications.

  • 178.
    Drabent, Wlodzimierz
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten. Institute of Computer Science, Polish Academy of Sciences.
    On definite program answers and least Herbrand models2016Ingår i: Theory and Practice of Logic Programming, ISSN 1471-0684, E-ISSN 1475-3081, Vol. 16, nr 4, 498-508 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    A sufficient and necessary condition is given under which least Herbrand models exactlycharacterize the answers of definite clause programs.

  • 179.
    Drabent, Wlodzimierz
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten. Polish Academic Science, Poland.
    Proving completeness of logic programs with the cut2017Ingår i: Formal Aspects of Computing, ISSN 0934-5043, E-ISSN 1433-299X, Vol. 29, nr 1, 155-172 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    Completeness of a logic program means that the program produces all the answers required by its specification. The cut is an important construct of programming language Prolog. It prunes part of the search space, this may result in a loss of completeness. This paper proposes a way of proving completeness of programs with the cut. The semantics of the cut is formalized by describing how SLD-trees are pruned. A sufficient condition for completeness is presented, proved sound, and illustrated by examples.

  • 180.
    Drabent, Włodzimierz
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten. Institute of Computer Science, Polish Academy of Sciences, Warzawa, Poland.
    Correctness and Completeness of Logic Programs2016Ingår i: ACM Transactions on Computational Logic, ISSN 1529-3785, E-ISSN 1557-945X, Vol. 17, nr 3, 18Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    We discuss proving correctness and completeness of definite clause logic programs.  We propose a method for proving completeness, while for proving correctness we employ a method which should be well known but is often neglected.  Also, we show how to prove completeness and correctness in the presence of SLD-tree pruning, and point out that approximate specifications simplify specifications and proofs.

    We compare the proof methods to declarative diagnosis (algorithmic debugging), showing that approximate specifications eliminate a major drawback of the latter.  We argue that our proof methods reflect natural declarative thinking about programs, and that they can be used, formally or informally, in every-day programming.

  • 181.
    Durairaj, Selva Ganesh
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Parallelize Automated Tests in a Build and Test Environment2016Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    This thesis investigates the possibilities of finding solutions, in order to reduce the total time spent for testing and waiting times for running multiple automated test cases in a test framework. The “Automated Test Framework”, developed by Axis Communications AB, is used to write the functional tests to test both hardware and software of a resource. The functional tests that tests the software is considered in this thesis work. In the current infrastructure, tests are executed sequentially and resources are allocated using First In First Out scheduling algorithm. From the user’s point of view, it is inefficient to wait for many hours to run their tests that take few minutes to execute. The thesis consists of two main parts: (1) identify a plugin that suits the framework and executes the tests in parallel, which reduces the overall execution time of tests and (2) analyze various scheduling algorithms in order to address the resource allocation problem, which arose due to limited resource availability, while the tests were run in parallel. By distributing multiple tests across several resources and executing them in parallel, help in improving the test strategy, thereby reducing the overall execution times of test suites. The case studies were created to emulate the problematic scenarios in the company and sample tests were written that reflect the real tests in the framework. Due to the complexity of the current architecture and the limited resources available for running the test in parallel, a simulator was developed with the identified plugin in a multi-core computer, with each core simulating a resource. Multiple tests were run using the simulator in order to explore, check and assess if the overall execution time of the tests can be reduced. While achieving parallelism in running the automated tests, resource allocation became a problem, since limited resources are available to run parallel tests. In order to address this problem, scheduling algorithms were considered. A prototype was developed to mimic the behaviour of a scheduling plugin and the scheduling algorithms were implemented in the prototype. The set of values were given as input to the prototype and tested with scenarios described under case studies. The results from the prototype are used to analyze the impact caused by various scheduling algorithms on reducing the waiting times of the tests. The combined usage of simulator along with scheduler prototype helped in understanding how to minimize the total time spent for testing and improving the resource allocation process.

  • 182.
    Eilert, Rickard
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Development of a framework for creating cross-platform TV HTML5 applications2015Självständigt arbete på grundnivå (yrkesexamen), 10,5 poäng / 16 hpStudentuppsats (Examensarbete)
    Abstract [sv]

    När man utvecklar HTML5-applikationer för TV-plattformar, finns utöver HTML5s standardfunktionalitet,även extra APIer för att komma åt TV-specifika funktioner. Dessa extra APIer skiljer sig åt mellan TV-plattformar, och det är ett problem när man utvecklar en applikation för många plattformar. Dettaexamensarbete har undersökt om det är möjligt att designa ett ramverk som förser utvecklaren med ett APIsom fungerar för många plattformar genom att wrappa deras plattformsspecifika kod. Svaret är ja. Ramverketharmoniserar: fjärrkontrollsindata, video, volym, internetförbindelse, TV-kanalströmmar och deras TV-tablåer, till ett gemensamt JavaScript-bibliotek-API. Vidare paketerar ett byggsystem koden på de sätt somplattformarna förväntar sig. Ramverket underlättar utveckling av HTML5-applikationer för TV-plattformar.För tillfället stöder ramverket plattformarna: Pace, PC och Samsung Smart TV, men det kan breddas med flerTV-plattformar.

  • 183.
    Einarson, Carl
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    An extension of the PPSZ Algorithm to Infinite-Domain Constraint Satisfaction Problems2017Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    The PPSZ algorithm (Paturi et al., FOCS 1998) is the fastest known algorithm for solving k-SAT when k >= 4. Hertli et al. recently extended the algorithm to solve (d, k)-Clause Satisfaction problems ((d,k)-ClSP) for which it is the fastest known algorithm for all k >= 3 (Hertli et al. CP 2016). We analyze their algorithm and extend it to solve problems over an infinite domain. More specifically we show how the extended algorithm can solve problems that have an infinite domain but where we can, for each instance of the problem, find a finite subset of the domain which has the following properties: If there exists a solution to the problem instance, then there exists a solution using only values from this subset and the size of this subset is polynomial in the size of the problem instance. We show numerically that our algorithm is the fastest known for problems over bounded disjunction languages for some values of k <= 500 and we look at the branching time temporal language, which is a bounded disjunction language, to show how to transform a specific problem to (d,k)-ClSP. We also look at Allen's interval algebra but conclude that there is already a faster algorithm for solving this problem.

  • 184.
    Enblom, Gustav
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    Eskebaek, Hannes
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    Real Time Vehicle Diagnostics Using Head Mounted Displays2015Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    This thesis evaluates how a head mounted display (HMD) can be used to increase usability compared to existing computer programs that are used during maintenance work on vehicles. Problems identified during a case study in a vehicle workshop are first described. As an attempt to solve some of the identified problems a prototype application using a HMD was developed. The prototype application aids the user during troubleshooting of systems on the vehicle by leading the mechanic with textual information and augmented reality (AR). Assessment of the prototype application was done by comparing it to the existing computer program and measuring error rate and time to completion for a predefined task. Usability was also measured using the System Usability Scale. The assessment showed that HMDs can provide higher usability in terms of efficiency and satisfaction. Furthermore, the thesis describes and discusses other possibilities and limitations that usage of HMDs and AR can lead to that were identified both from theory and during implementation.

  • 185.
    Englund, Albin
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Suther, Magnus
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Bluetooth Low Energy som trådlös standard för hemautomation2013Självständigt arbete på grundnivå (kandidatexamen), 10 poäng / 15 hpStudentuppsats (Examensarbete)
    Abstract [sv]

    Allmänheten har en stor efterfrågan av produkter inom området för hemautomation. Den senaste bluetoothstandarden Bluetooth Low Energy skapar nya möjligheter för intressanta produkter som underlättar vardagen. Lösningar som IR och Wi-Fi ger inte de förutsättningar som krävs för att på ett energisnålt och praktiskt sätt erbjuda sådana produkter, något som Bluetooth Low Energy gör. I denna rapport diskuteras standarden i syfte att redogöra för hur den kan användas för att automatisera ett hem.

    För detta examensarbete implementerades en strömbrytarprototyp och en iOS-applikation, vilka användes för att unders öka och påvisa ett koncept för hur tekniken kan tillämpas för hemautomation. Resultaten visar att teknikens räckvidd är dess främsta begränsning. Likaså visas hur signalstyrkan kan användas som en utlösande faktor för att styra en strömbrytare.

    Denna rapport redogör också för hur systemet uppnår interoperabilitet genom att implementera en utarbetad profil.

  • 186. Engstrom, Robert
    et al.
    Färnqvist, Tommy
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    Jonsson, Peter
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    Thapper, Johan
    University of Paris Est Marne La Vallee, France.
    An Approximability-related Parameter on Graphs - Properties and Applications2015Ingår i: DISCRETE MATHEMATICS AND THEORETICAL COMPUTER SCIENCE, ISSN 1462-7264, Vol. 17, nr 1, 33-66 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    We introduce a binary parameter on optimisation problems called separation. The parameter is used to relate the approximation ratios of different optimisation problems; in other words, we can convert approximability (and non-approximability) result for one problem into (non)-approximability results for other problems. Our main application is the problem (weighted) maximum H-colourable subgraph (MAX H-COL), which is a restriction of the general maximum constraint satisfaction problem (MAX CSP) to a single, binary, and symmetric relation. Using known approximation ratios for MAX k-CUT, we obtain general asymptotic approximability results for MAX H-COL for an arbitrary graph H. For several classes of graphs, we provide near-optimal results under the unique games conjecture. We also investigate separation as a graph parameter. In this vein, we study its properties on circular complete graphs. Furthermore, we establish a close connection to work by Samal on cubical colourings of graphs. This connection shows that our parameter is closely related to a special type of chromatic number. We believe that this insight may turn out to be crucial for understanding the behaviour of the parameter, and in the longer term, for understanding the approximability of optimisation problems such as MAX H-COL.

  • 187.
    Engström, Adam
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Strömsparande arkitektur för inbyggnadslinux2014Självständigt arbete på grundnivå (kandidatexamen), 10 poäng / 15 hpStudentuppsats (Examensarbete)
    Abstract [en]

    The objective of this work was to evaluate and implement a number of energy saving functions for a specific embedded system. The functions were then grouped into a number of energy levels with known properties in terms of functionality, energy consumption, and transition time between the levels.

    The embedded system consisted of an AT91 ARM9 processor, GSM/GPRS modem, display, Ethernet and other peripheral units. Some energy saving methods that were considered were suspend to RAM, suspend to disk, frequency scaling, and methods for saving energy in the modem, Ethernet, USB and display backlight. The functions were grouped into levels and an interface was specified for controlling the energy level.

    It proved possible to get known properties within the defined energy levels, even though the paritioning of functions into these levels proved to be sub-optimal in a typical application usage scenario because it was designed for mainly energy consumption, not usage.

    The final result is a number of energy saving functions grouped into levels, which are controllable via an application interface. Each of the levels have a known energy consumption in both loaded and un-loaded mode.

  • 188.
    Eriksson, Björn
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    A study of Bitcoin as a currency for email-based micro-transactions2016Självständigt arbete på grundnivå (kandidatexamen), 10,5 poäng / 16 hpStudentuppsats (Examensarbete)
    Abstract [en]

    Bitcoin is a cryptocurrency that has been the focus of a lot of discussions lately and has attracted a large number of users. Its offers many possibilities for cheap transactions and unregulated finances which has been realized in numerous sites and applications on the web and in mobile phones. One medium that seem to have been neglected when it comes to Bitcoins development is email. This is curious since Bitcoin by its nature seem to have many properties that would work well with texted messages. The purpose of this study it to analyze the current papers about Bitcoin to find the current status of email based Bitcoin services,  and try to analyze if email as a tool is a suitable medium to be used with Bitcoin. This analyze is done through a systematic literature review of current papers, followed by an examination of past and current Bitcoin companies that has used email as part of their service. In the end the results suggested that the low security in email, and the apparent lack of services that would benefit from an email based Bitcoin service suggest that a service like that would be hard to develop today and not very useful to the public.

  • 189.
    Eriksson, Joakim
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Representation of asynchronous communication protocols in Scala and Akka2013Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    This thesis work investigates how to represent protocols for asynchronous communication in the Scala programming language and the Akka actor framework, to be run on Java Virtual Machine (JVM). Further restrictions from the problem domain - the coexistence of multiple protocol instances sharing the same Java thread - imply that neither an asynchronous call waiting for response nor anything else can block the underlying Java threads.

    A common way to represent asynchronous communication protocols is to use state machines. This thesis seeks a way to shrink the size of and to reduce the complexity of the protocol implementations by representing sequences of asynchronous communication calls (i.e. sequences of sent and received messages) as a type of procedure. The idea is find a way to make the procedures that contain asynchronous calls look like synchronous communication procedures by hiding the asynchronous details. In other words, the resulting procedure code should show what to do and not so much focus on how to overcome the impediment of the asynchronous calls.

    With the help of an asynchronous communication protocol toy example, this report shows how such an protocol can be implemented with a combination of a state machine and a procedure representation in Scala and Akka. The procedure representation hides away the asynchronous details by using the Scala capability to use CPS-transformed delimited continuations. As a sub-problem, this thesis also shows how to safely schedule asynchronous communication timeouts with help of Scala and Akka within the restrictions of the thesis problem domain.

  • 190.
    Eriksson, Jonas
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    Partitioning methodology validation for embedded systems design2016Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    As modern embedded systems are becoming more sophisticated the demands on their applications significantly increase. A current trend is to utilize the advances of heterogeneous platforms (i.e. platform consisting of different computational units (e.g. CPU, FPGA or GPU)) where different parts of the application can be distributed among the different computational units as software and hardware implementations. This technology can improve the application characteristics to meet requirements (e.g. execution time, power consumption and design cost), but it leads to a new challenge in finding the best combination of hardware and software implementation (referred as system configuration). The decisions whether a part of the application should be implemented in software (e.g. as C code) or hardware (e.g. as VHDL code) affect the entire product life-cycle. This is traditionally done manually by the developers in the early stage of the design phase. However, due to the increasing complexity of the application the need of a systematic process that aids the developer when making these decisions to meet the demands rises. Prior to this work a methodology called MULTIPAR has been designed to address this problem. MULTIPAR applies component-/model-based techniques to design the application, i.e. the application is modeled as a number of interconnected components, where some of the components will be implemented as software and the remaining ones as hardware. To perform the partitioning decisions, i.e. determining for each component whether it should be implemented as software or hardware, MULTIPAR proposes a set of formulas to calculate the properties of the entire system based on the properties for each component working in isolation.

    This thesis aims to show to what extent the proposed system formulas are valid. In particular it focuses on validating the formulas that calculate the system response time, system power consumption, system static memory and system FPGA area. The formulas were validated trough an industrial case study, where the system properties for different system configurations were measured and calculated by applying these formulas. The measured values and calculated values for the system properties were compared by conducting a statistical analysis. The case study demonstrated that the system properties can be accurately calculated by applying the system formulas.

  • 191.
    Eriksson, Mattias
    et al.
    Linköpings universitet, Institutionen för datavetenskap, PELAB - Laboratoriet för programmeringsomgivningar. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Integrated Code Generation for Loops2012Ingår i: ACM Transactions on Embedded Computing Systems, ISSN 1539-9087, E-ISSN 1558-3465, Vol. 11, nr 1Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    Code generation in a compiler is commonly divided into several phases: instruction selection, scheduling, register allocation, spill code generation, and, in the case of clustered architectures, cluster assignment. These phases are interdependent; for instance, a decision in the instruction selection phase affects how an operation can be scheduled We examine the effect of this separation of phases on the quality of the generated code. To study this we have formulated optimal methods for code generation with integer linear programming; first for acyclic code and then we extend this method to modulo scheduling of loops. In our experiments we compare optimal modulo scheduling, where all phases are integrated, to modulo scheduling, where instruction selection and cluster assignment are done in a separate phase. The results show that, for an architecture with two clusters, the integrated method finds a better solution than the nonintegrated method for 27% of the instances.

  • 192.
    Eriksson, Mattias
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Integrated Offset Assignment2011Ingår i: 9th Workshop on Optimizations for DSP and Embedded Systems (ODES-9) / [ed] George Cai, Tom van der Aa, odes-workshop.weebly.com , 2011, 47-54 s.Konferensbidrag (Refereegranskat)
    Abstract [en]

    One important part of generating code for DSP processors is to make good use of the address generation unit (AGU). In this paper we divide the code generation into three parts: (1) scheduling, (2) address register assignment, and (3) storage layout. The goal is to nd out if solving these three subproblems as one big integrated problem gives better results compared to when scheduling or address register assignment is solved separately. We present optimal dynamic programming algorithms for both integrated and non-integrated code generation for DSP processors. In our experiments we nd that integrationis benecial when the AGU has 1 or 2 address registers; for the other cases existing heuristics are near optimal. We also nd that integrating address register assignment and storage layout gives slightly better results than integrating scheduling and storage layout. I.e. address register assignment is more important than scheduling.

  • 193.
    Eriksson, Oskar
    et al.
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    Rydkvist, Emil
    Linköpings universitet, Institutionen för teknik och naturvetenskap, Medie- och Informationsteknik. Linköpings universitet, Tekniska fakulteten.
    An in-depth analysis of dynamically rendered vector-based maps with WebGL using Mapbox GL JS2015Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    The regular way of displaying maps in a web browser is by downloading raster images from a server and lay them side by side to make up a map. If any information on the map is changed, new images has to be downloaded, it cannot be done on the client. The introduction of WebGL opens up a whole new world of delivering advanced graphics content to the end user in a web browser. Utilizing this technology for displaying maps means only the source data is sent to the web browser where the map gets rendered using the device's GPU. This adds a number of benefits such as the ability of changing map appearance on the client, add new features to the map and often less data transfer. It however sets higher expectations of the client device's hardware as it needs to render the map at a high enough frame rate to not appear slow and unresponsive. This thesis investigates a framework for client side map rendering in a web browser, Mapbox GL JS, with focus on performance. It shows how map source data can be generated as well as its corresponding style rules are constructed with performance in mind. It provides benchmarking results of different map data sets with different detail intensity and shows that a device with good GPU performance is needed for an acceptable user experience. It also shows that lowering the amount of rendered detail does not necessarily result in better performance.

  • 194.
    Ernstsson, August
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    SkePU 2: Language Embedding and Compiler Support for Flexible and Type-Safe Skeleton Programming2016Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    This thesis presents SkePU 2, the next generation of the SkePU C++ framework for programming of heterogeneous parallel systems using the skeleton programming concept. SkePU 2 is presented after a thorough study of the state of parallel programming models, frameworks and tools, including other skeleton programming systems. The advancements in SkePU 2 include a modern C++11 foundation, a native syntax for skeleton parameterization with user functions, and an entirely new source-to-source translator based on Clang compiler front-end libraries.

    SkePU 2 extends the functionality of SkePU 1 by embracing metaprogramming techniques and C++11 features, such as variadic templates and lambda expressions. The results are improved programmability and performance in many situations, as shown in both a usability survey and performance evaluations on high-performance computing hardware. SkePU’s skeleton programming model is also extended with a new construct, Call, unique in the sense that it does not impose any predefined skeleton structure and can encapsulate arbitrary user-defined multi-backend computations.

    We conclude that SkePU 2 is a promising new direction for the SkePU project, and a solid basis for future work, for example in performance optimization.

  • 195.
    Faur, Andrei
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Memory Profiling Techniques2012Självständigt arbete på avancerad nivå (masterexamen), 20 poäng / 30 hpStudentuppsats (Examensarbete)
    Abstract [en]

    Memory profiling is an important technique which aids program optimization and can even help tracking down bugs. The main problem with the current memory profiling techniques and tools is that they slow down the target software considerably therefore making them inadequate for mainline integration. Ideally, the user would be able to monitor memory consumption without having to worry about the rest of the software being affected in any way. This thesis provides a comparison of existing techniques and tools along with the description of a memory profiler implementation which tries to provide a balance between the information it is able to retrieve and the influence it has on the target software.

  • 196.
    Ferretti, Gianni
    et al.
    Politecn Milan, Italy.
    Fritzson, Peter
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska fakulteten.
    Editorial Material: Special issue on object-oriented modelling and simulation in MATHEMATICAL AND COMPUTER MODELLING OF DYNAMICAL SYSTEMS, vol 23, issue 3, pp 240-2422017Ingår i: Mathematical and Computer Modelling of Dynamical Systems, ISSN 1387-3954, E-ISSN 1744-5051, Vol. 23, nr 3, 240-242 s.Artikel i tidskrift (Övrigt vetenskapligt)
    Abstract [en]

    n/a

  • 197.
    Fors Johansson, Christoffer
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system.
    OpenModelica Interactive Simulation using an OPC UA client2017Självständigt arbete på grundnivå (högskoleexamen), 10,5 poäng / 16 hpStudentuppsats (Examensarbete)
    Abstract [en]

    Simulating a model of a complex physical system can beuseful in order to obtain a deeper understanding of the system.OpenModelica is an open-source, modeling and simulation environmentin which such a simulation could be performed. This thesispresents a design and implementation of an interactive simulationprototype for the OpenModelica Connection Editor. After aninvestigation of the possibilities in the current OpenModelicaenvironment, several requirements were determined about howinteractive simulation should look and work. A prototype wasdeveloped based on these requirements. As a result, the prototypecan act as a base for applying further functionality to theinteractive simulation in both parallel and future works.

  • 198.
    Forsell, Martti
    et al.
    Platform Architectures Team, VTT Technical Research Centre of Finland, Finland.
    Hansson, Erik
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Mäkelä, Jari-Matti
    Information Technology, University of Turku, Finland.
    Leppänen, Ville
    Information Technology, University of Turku, Finland.
    Hardware and Software Support for NUMA Computing on Configurable Emulated Shared Memory Architectures2013Ingår i: 2013 IEEE 27th International Parallel and Distributed Processing Symposium Workshops & PhD Forum (IPDPSW), IEEE conference proceedings, 2013, 640-647 s.Konferensbidrag (Refereegranskat)
    Abstract [en]

    The emulated shared memory (ESM) architectures are good candidates for future general purpose parallel computers due to their ability to provide easy-to-use explicitly parallel synchronous model of computation to programmers as well as avoid most performance bottlenecks present in current multicore architectures. In order to achieve full performance the applications must, however, have enough thread-level parallelism (TLP). To solve this problem, in our earlier work we have introduced a class of configurable emulated shared memory (CESM) machines that provides a special non-uniform memory access (NUMA) mode for situations where TLP is limited or for direct compatibility for legacy code sequential computing or NUMA mechanism. Unfortunately the earlier proposed CESM architecture does not integrate the different modes of the architecture well together e.g. by leaving the memories for different modes isolated and therefore the programming interface is non-integrated. In this paper we propose a number of hardware and software techniques to support NUMA computing in CESM architectures in a seamless way. The hardware techniques include three different NUMA-shared memory access mechanisms and the software ones provide a mechanism to integrate NUMA computation into the standard parallel random access machine (PRAM) operation of the CESM. The hardware techniques are evaluated on our REPLICA CESM architecture and compared to an ideal CESM machine making use of the proposed software techniques.

  • 199.
    Forsell, Martti
    et al.
    Platform Architectures Team, VTT Technical Research Centre of Finland, Finland.
    Hansson, Erik
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Kessler, Christoph
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan.
    Mäkelä, Jari-Matti
    Department of Information Technology, University of Turku, Finland.
    Leppänen, Ville
    Department of Information Technology, University of Turku, Finland.
    NUMA Computing with Hardware and Software Co-Support on Configurable Emulated Shared Memory Architectures2014Ingår i: International Journal of Networking and Computing, ISSN 2185-2839, E-ISSN 2185-2847, Vol. 4, nr 1, 189-206 s.Artikel i tidskrift (Refereegranskat)
    Abstract [en]

    The emulated shared memory (ESM) architectures are good candidates for future general purpose parallel computers due to their ability to provide an easy-to-use explicitly parallel synchronous model of computation to programmers as well as avoid most performance bottlenecks present in current multicore architectures. In order to achieve full performance the applications must, however, have enough thread-level parallelism (TLP). To solve this problem, in our earlier work we have introduced a class of configurable emulated shared memory (CESM) machines that provides a special non-uniform memory access (NUMA) mode for situations where TLP is limited or for direct compatibility for legacy code sequential computing and NUMA mechanism. Unfortunately the earlier proposed CESM architecture does not integrate the different modes of the architecture well together e.g. by leaving the memories for different modes isolated and therefore the programming interface is non-integrated. In this paper we propose a number of hardware and software techniques to support NUMA computing in CESM architectures in a seamless way. The hardware techniques include three different NUMA shared memory access mechanisms and the software ones provide a mechanism to integrate and optimize NUMA computation into the standard parallel random access machine (PRAM) operation of the CESM. The hardware techniques are evaluated on our REPLICA CESM architecture and compared to an ideal CESM machine making use of the proposed software techniques.

  • 200.
    Forsgren, Mikael
    et al.
    Linköpings universitet, Institutionen för medicin och hälsa, Avdelningen för radiologiska vetenskaper. Linköpings universitet, Hälsouniversitetet. Linköpings universitet, Centrum för medicinsk bildvetenskap och visualisering, CMIV. Östergötlands Läns Landsting, Centrum för kirurgi, ortopedi och cancervård, Radiofysikavdelningen US.
    Dahlström, Nils
    Linköpings universitet, Institutionen för medicin och hälsa, Avdelningen för radiologiska vetenskaper. Linköpings universitet, Hälsouniversitetet. Linköpings universitet, Centrum för medicinsk bildvetenskap och visualisering, CMIV. Östergötlands Läns Landsting, Diagnostikcentrum, Röntgenkliniken i Linköping.
    Karlsson, Markus
    Linköpings universitet, Institutionen för datavetenskap, Programvara och system. Linköpings universitet, Tekniska högskolan. Linköpings universitet, Institutionen för klinisk och experimentell medicin, Avdelningen för neurovetenskap.
    Dahlqvist Leinhard, Olof
    Linköpings universitet, Institutionen för medicin och hälsa, Avdelningen för radiologiska vetenskaper. Linköpings universitet, Hälsouniversitetet. Linköpings universitet, Centrum för medicinsk bildvetenskap och visualisering, CMIV. Östergötlands Läns Landsting, Centrum för kirurgi, ortopedi och cancervård, Radiofysikavdelningen US.
    Smedby, Örjan
    Linköpings universitet, Institutionen för medicin och hälsa, Avdelningen för radiologiska vetenskaper. Linköpings universitet, Hälsouniversitetet. Linköpings universitet, Centrum för medicinsk bildvetenskap och visualisering, CMIV. Östergötlands Läns Landsting, Diagnostikcentrum, Röntgenkliniken i Linköping.
    Cedersund, Gunnar
    Linköpings universitet, Institutionen för medicinsk teknik. Linköpings universitet, Hälsouniversitetet.
    Lundberg, Peter
    Linköpings universitet, Institutionen för medicin och hälsa, Avdelningen för radiologiska vetenskaper. Linköpings universitet, Hälsouniversitetet. Linköpings universitet, Centrum för medicinsk bildvetenskap och visualisering, CMIV. Östergötlands Läns Landsting, Centrum för kirurgi, ortopedi och cancervård, Radiofysikavdelningen US. Östergötlands Läns Landsting, Diagnostikcentrum, Röntgenkliniken i Linköping.
    Whole Body Mechanistic Minimal Model for Gd-EOB-DTPA Contrast Agent Pharmacokinetics in Evaluation of Diffuse Liver Disease2014Konferensbidrag (Övrigt vetenskapligt)
    Abstract [en]

    Purpose: Aiming for non-invasive diagnostic tools to decrease the need for biopsy in diffuse liver disease and to quantitatively describe liver function, we applied a mechanistic pharmacokinetic modelling analysis of liver MRI with Gd-EOB-DTPA. This modelling method yields physiologically relevant parameters and was compared to previously developed methods in a patient group with diffuse liver disease. Materials and Methods: Using data from healthy volunteers undergoing liver MRI, an identifiable mechanistic model was developed, based on compartments described by ordinary differential equations and kinetic expressions, and validated with independent data including Gd-EOB-DTPA concentration measurements in blood samples. Patients (n=37) with diffuse liver disease underwent liver biopsy and MRI with Gd-EOB-DTPA. The model was used to derive pharmacokinetic parameters which were then compared with other quantitative estimates in their ability to separate mild from severe liver fibrosis. Results: The estimations produced by the mechanistic model allowed better separation between mild and severe fibrosis than previously described methods for quantifying hepatic Gd-EOB-DTPA uptake. Conclusions: With a mechanistic pharmacokinetic modelling approach, the estimation of liver uptake function and its diagnostic information can be improved compared to current methods.

1234567 151 - 200 av 534
RefereraExporteraLänk till träfflistan
Permanent länk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf