Over Sundvollen
NIKT 2020 at USN, logo

24th - 25th November

Norwegian Informatics Conference

University of South-Eastern Norway

Feasibility of Optimizations Requiring Bounded Treewidth in a Data Flow Centric Intermediate Representation
Sigve Sjømæling Nordgaard and Jan Christian Meyer Department of Computer Science, NTNU Trondheim, Norway
Data flow analyses are instrumental to effective compiler optimizations, and are typically implemented by extracting implicit data flow information from traversals of a control flow graph intermediate representation. The Regionalized Value State Dependence Graph is an alternative intermediate representation, which represents a program in terms of its data flow dependencies, leaving control flow implicit. Several analyses that enable compiler optimizations reduce to NP-Complete graph problems in general, but admit linear time solutions if the graph’s treewidth is limited. In this paper, we investigate the treewidth of application benchmarks and synthetic programs, in order to identify program features which cause the treewidth of its data flow graph to increase, and assess how they may appear in practical software. We find that increasing numbers of live variables cause unbounded growth in data flow graph treewidth, but this can ordinarily be remedied by modular program design, and monolithic programs that exceed a given bound can be efficiently detected using an approximate treewidth heuristic.
Evaluating multi-core graph algorithm frameworks
Zawadi Svela
Department of Computer Science Norwegian University of Science and Technology z.b.svela@gmail.com
Multi-core and GPU-based systems offer unprecedented computational power. They are, however, challenging to utilize effectively, especially when processing irregular data such as graphs. Graphs are of great interest, as they are now used to model geographic-, social- and neural networks. Several interesting programming frameworks for graph processing have therefore been developed these past few years.
In this work, we highlight the strengths and weaknesses of the Galois, GraphBLAST, Gunrock and Ligra graph frameworks through benchmarking their single source shortest path (SSSP) implementations using the SuiteSparse Matrix Collection. Tests were done on an Nvidia DGX2 system, except for Ligra, which only provides a multi-core framework. D-IrGL, built on Galois, also provided a multi-GPU option for SSSP. We also look at program size, documentation and overall ease of use.
High performance generally comes at the price of high complexity. D-IrGL shows its strength on the very largest graphs, where it achieved the best run-time, while Gunrock processed most other large sets the fastest. However, GraphBLAST, with a relatively low- complexity interface, achieves the greatest median throughput across all our test cases. This despite that its SSSP implementation size is only 1/10th of Gunrock, which for our tests has the highest peak throughput and the fastest run-time in most cases. Ligra had less computational resources available, and consequently performed worse in most cases, but it is also a very compact and easy to use framework. Futher analyses and some suggestions for future work are also included.
Some Faster Algorithms for Finding Large Prime Gaps
Arne Maus (em), Thomas Gulli and Eric B. Jul
(arnem,thomgull,ericbj@ifi.uio.no) Programming Technology Group
Dept. of Informatics, University of Oslo, Norway
This paper investigates the problem of finding large prime gaps (the difference between two consecutive prime numbers, pi+1 – pi) and on the development of a small, efficient program for generating such large prime gaps for a single computer, a laptop or a workstation.
Parallel Scalability of Adaptive Mesh Refinement in a Finite Difference Solution to the Shallow Water Equations
Ola Toft and Jan Christian Meyer Department of Computer Science, NTNU Trondheim, Norway
The Shallow Water Equations model the fluid dynamics of deep ocean flow, and are used to simulate tides, tsunamis, and storm surges. Numerical solutions using finite difference methods are computationally expensive enough to mandate the use of large computing clusters, and the cost grows not only with the amount of fluid, but also the duration of the simulated event, and the resolution of the approximation. The benefits of increased resolution are mostly connected to regions where complex fluid interactions occur, and are not required globally for the entire simulation. In this paper, we investigate the potential for conserving computational resources by applying Adaptive Mesh Refinement to dynamically determined areas of the fluid surface. We implement adaptive mesh refinement in a MacCormack finite difference solver, develop a performance model to predict its behavior on large-scale parallel platforms, and validate its predictions experimentally on two computing clusters. We find that the solver itself has highly favorable parallel scalability, and that the addition of refined areas introduces a performance penalty due to load imbalance that is at most proportional to the refinement degree raised to the third power.
Addressing the ethical principles of the Norwegian National Strategy for AI in a kindergarten allocation system
Stian Botnevik Andrey Belinskiy Edis Asotic Herman Stoud Platou Thomas Søvik Marija Slavkovik
The Norwegian National Strategy for Artificial Intelligence (NNSAI) published in 2020 includes seven principles of ethical AI. This paper explores whether those seven principles are stated in a clear enough way and are feasible to be satisfied by a specific AI system. We build an implementation of the Gale-Shapley algorithm to allocate kindergarten places in Bergen, Norway. The presented solution is then evaluated against the ethical principles from the NNSAI. We argue that it is difficult to respect all the ethical principles when implementing a solution to a matching problem.
Bias mitigation with AIF360: A comparative study
TorH.Aasheim KnutT.Hufthammer SølveA ̊nneland Ha ̊vard Brynjulfsen Marija Slavkovik
University of Bergen
The use of artificial intelligence for decision making raises concerns about the societal impact of such systems. Traditionally, the product of a human decision-maker are governed by laws and human values. Decision-making is now being guided - or in some cases, replaced by machine learning classification which may reinforce and introduce bias. Algorithmic bias mitigation is explored as an approach to avoid this, however it does come at a cost: efficiency and accuracy. We conduct an empirical analysis of two off-the-shelf bias mitigation techniques from the AIF360 toolkit on a binary classification task. Our preliminary results indicate that bias mitigation is a feasible approach to ensuring group fairness.
Deep Active Learning for Autonomous Perception
Navjot Singh, Håkon Hukkelås and Frank Lindseth
Traditional supervised learning requires significant amounts of labeled training data to achieve satisfactory results. As autonomous perception systems collect continuous data, the labeling process becomes expensive and time-consuming. Active learning is a specialized semi- supervised learning strategy that allows a machine learning model to achieve high performance using less training data, thereby minimizing the cost of manual annotation. We explore active learning for autonomous vehicles, and propose a novel deep active learning framework for object detection and instance segmentation. We review prominent active learning approaches, study their performances in the aforementioned computer vision tasks, and perform several experiments using state-of-the-art R-CNN-based models for datasets in the self-driving domain. Our empirical experiments on a number of datasets reflect that active learning reduces the amount of training data required. We observe that early exploration with instance- rich training sets leads to good performance, and that false positives can have a negative impact if not dealt with appropriately. Furthermore, we perform a qualitative evaluation using autonomous driving data collected from Trondheim, illustrating that active learning can help in selecting more informative images to annotate.
Enhancing Bi-directional English-Tigrigna Machine Translation Using Hybrid Approach
Zemicheal Berihu1, Gebremariam Mesfin, Mulugeta Atsibaha, Tor-Morten Grønli
Aksum University, Department of Computing Technology, Aksum, Ethiopia Kristiania University College, Department of Technology, Oslo, Norway
Machine Translation (MT) is an application area of NLP where automatic systems are used to translate text or speech from one language to another while preserving the meaning of the source language. Although there exists a large volume of literature in automatic machine translation of documents in many languages, the translation between English and Tigrigna is less explored. Therefore, we proposed the hybrid approach to address the challenges of applying syntactic reordering rules which align and capture the structural arrangement of words in the source sentence to become more like the target sentences. Two language models were developed- one for English and another for Tigrigna and about 12,000 parallel sentences in four domains and 32,000 bilingual dictionaries were collected for our experiment. The parallel collected corpus was split randomly to 10,800 sentences for training set and 1,200 sentences for testing. Moses open source statistical machine translation system has been used for the experiment to train, tune and decode. The parallel corpus was aligned using the Giza++ toolkit and SRILM was used for building the language model. Three main experiments were conducted using statistical approach, hybrid approach and post-processing technique. According to our experimental result showed good translation output as high as 32.64 BLEU points Google translator and the hybrid approach was found most promising for English-Tigrigna bi-directional translation.
2D and 3D U-Nets for skull stripping in a large and heterogeneous set of head MRI using fastai
Satheshkumar Kaliyugarasan, Marek Kocinski, Arvid Lundervold, Alexander Selvikvåg Lundervold
Skull stripping in brain imaging is the removal of the parts of images corresponding to non-brain tissue. Fast and accurate skull stripping is a crucial step for numerous medical brain imaging applications, e.g. registration, segmentation and feature extraction, as it eases subsequent image processing steps. In this work, we propose and compare two novel skull stripping methods based on 2D and 3D convolutional neural networks trained on a large, heterogeneous collection of 2777 clinical 3D T1-weighted MRI images from 1681 healthy subjects. We investigated the performance of the models by testing them on 927 images from 324 subjects set aside from our collection of data, in addition to images from an independent, large brain imaging study: the IXI dataset (n = 556). Our models achieved mean Dice scores higher than 0.978 and Jaccard indices higher than 0.957 on all tests sets, making predictions on new unseen brain MR images in approximately 1.4s for the 3D model and 12.4s for the 2D model. A preliminary exploration of the models’ robustness to variation in the input data showed favourable results when compared to a traditional, well-established skull stripping method. With further research aimed at increasing the models’ robustness, such accurate and fast skull stripping methods can potentially form a useful component of brain MRI analysis pipelines.
Synthesizing skin lesion images using CycleGANs – a case study
Sondre Fossen-Romsaas, Adrian Storm-Johannessen, and Alexander Selvikvåg Lundervold1
Generative adversarial networks (GANs) have seen some success as a way to synthesize training data for supervised machine learning models. In this work, we design two novel approaches for synthetic image generation based on CycleGANs, aimed at generating realistic-looking, class-specific dermoscopic skin lesion images. We evaluate the images’ usefulness as additional training data for a convolutional neural network trained to perform a difficult lesion classification task. We are able to generate visually striking images, but their value for augmenting the classifier’s training data set is low. This is in-line with other researcher’s investigations into similar GAN models, indicating the need for further research into forcing GAN models to produce samples further from the training data distribution, and to find ways of guiding the image generation using feedback from the ultimate classification objective.
The Role of Virtual Simulation in Incident Commander Education – A field study
Cecilia Hammar Wijkmark, Maria Monika Metallinou, Ilona Heldal Western Norway University of Applied Sciences, Norway, Sune Fankvist, The Swedish Civil Contingencies Agency, Sweden
The use of Virtual Simulation (VS) for emergency management and Incident Commander (IC) training and assessment has spread during the last decade. In VS, ICs act in computer- simulated 3D incident scenarios, e.g. fire incidents, road traffic collisions etc. Even though VS provides several benefits, there is a history of hesitation to implement and apply it in emergency education. This paper presents the results of a field study performed during the VS training in four classes of IC-students (90 students). The research focus was on the IC students` attitudes and experiences of VS training. Data were collected through observations and post-training questionnaires. The results show that students are predominantly positive towards virtual simulation. 72% of the IC-students state that they experienced presence to the same extent as in live simulation settings, where they experience high presence. Earlier, photorealism was considered to be necessary to provide virtual learning places with high experiences. According to this study, this is not equally important on a general base. The results argue for the benefits of using VS in IC training, even if there are challenges with the implementation. Furthermore, it contributes to a better understanding of user experiences and realism in VS training compared to live simulation.
Specifying Software Languages: Grammars, Projectional Editors, and Unconventional Approaches
Mikhail Barash, University of Bergen, Norway
We discuss several approaches for defining software languages, together with Integrated Development Environments for them. Theoretical foundation is grammar-based models: they can be used where proven correctness of specifications is required. From a practical point of view, we discuss how language specification can be made more accessible by focusing on language workbenches and projectional editing, and discuss how it can be formalized. We also give a brief overview of unconventional ideas to language definition, and outline three open problems connected to the approaches we discuss.