Skip to main content
Login | Suomeksi | På svenska | In English

Browsing by study line "ingen studieinriktning"

Sort by: Order: Results:

  • Kovanen, Veikko (2020)
    Real estate appraisal, or property valuation, requires strong expertise in order to be performed successfully, thus being a costly process to produce. However, with structured data on historical transactions, the use of machine learning (ML) enables automated, data-driven valuation which is instant, virtually costless and potentially more objective compared to traditional methods. Yet, fully ML-based appraisal is not widely used in real business applications, as the existing solutions are not sufficiently accurate and reliable. In this study, we introduce an interpretable ML model for real estate appraisal using hierarchical linear modelling (HLM). The model is learned and tested with an empirical dataset of apartment transactions in the Helsinki area, collected during the past decade. As a result, we introduce a model which has competitive predictive performance, while being simultaneously explainable and reliable. The main outcome of this study is the observation that hierarchical linear modelling is a very potential approach for automated real estate appraisal. The key advantage of HLM over alternative learning algorithms is its balance of performance and simplicity: this algorithm is complex enough to avoid underfitting but simple enough to be interpretable and easy to productize. Particularly, the ability of these models to output complete probability distributions quantifying the uncertainty of the estimates make them suitable for actual business use cases where high reliability is required.
  • Lintunen, Milla (2023)
    Fault management in mobile networks is required for detecting, analysing, and fixing problems appearing in the mobile network. When a large problem appears in the mobile network, multiple alarms are generated from the network elements. Traditionally Network Operations Center (NOC) process the reported failures, create trouble tickets for problems, and perform a root cause analysis. However, alarms do not reveal the root cause of the failure, and the correlation of alarms is often complicated to determine. If the network operator can correlate alarms and manage clustered groups of alarms instead of separate ones, it saves costs, preserves the availability of the mobile network, and improves the quality of service. Operators may have several electricity providers and the network topology is not correlated with the electricity topology. Additionally, network sites and other network elements are not evenly distributed across the network. Hence, we investigate the suitability of a density-based clustering methods to detect mass outages and perform alarm correlation to reduce the amount of created trouble tickets. This thesis focuses on assisting the root cause analysis and detecting correlated power and transmission failures in the mobile network. We implement a Mass Outage Detection Service and form a custom density-based algorithm. Our service performs alarm correlation and creates clusters of possible power and transmission mass outage alarms. We have filed a patent application based on the work done in this thesis. Our results show that we are able to detect mass outages in real time from the data streams. The results also show that detected clusters reduce the number of created trouble tickets and help reduce of the costs of running the network. The number of trouble tickets decreases by 4.7-9.3% for the alarms we process in the service in the tested networks. When we consider only alarms included in the mass outage groups, the reduction is over 75%. Therefore continuing to use, test, and develop implemented Mass Outage Detection Service is beneficial for operators and automated NOC.
  • Mäki, Niklas (2023)
    Most graph neural network architectures take the input graph as granted and do not assign any uncertainty to its structure. In real life, however, data is often noisy and may contain incorrect edges or exclude true edges. Bayesian methods, which consider the input graph as a sample from a distribution, have not been deeply researched, and most existing research only tests the methods on small benchmark datasets such as citation graphs. As often is the case with Bayesian methods, they do not scale well for large datasets. The goal of this thesis is to research different Bayesian graph neural network architectures for semi-supervised node classification and test them on larger datasets, trying to find a method that improves the baseline model and is scalable enough to be used with graphs of tens of thousands of nodes with acceptable latency. All the tests are done twice with different amounts of training data, since Bayesian methods often excel with low amounts of data and in real life labeled data can be scarce. The Bayesian models considered are based on the graph convolutional network, which is also used as the baseline model for comparison. This thesis finds that the impressive performance of the Bayesian graph neural networks does not generalize to all datasets, and that the existing research relies too much on the same small benchmark graphs. Still, the models may be beneficial in some cases, and some of them are quite scalable and could be used even with moderately large graphs.
  • Paulamäki, Henri (2019)
    Tailoring a hybrid surface or any complex material to have functional properties that meet the needs of an advanced device or drug requires knowledge and control of the atomic level structure of the material. The atomistic configuration can often be the decisive factor in whether the device works as intended, because the materials' macroscopic properties - such as electrical and thermal conductivity - stem from the atomic level. However, such systems are difficult to study experimentally and have so far been infeasible to study computationally due to costly simulations. I describe the theory and practical implementation of a 'building block'-based Bayesian Optimization Structure Search (BOSS) method to efficiently address heterogeneous interface optimization problems. This machine learning method is based on accelerating the identification of a material's energy landscape with respect to the number of quantum mechanical (QM) simulations executed. The acceleration is realized by applying likelihood-free Bayesian inference scheme to evolve a Gaussian process (GP) surrogate model of the target landscape. During this active learning, various atomic configurations are iteratively sampled by running static QM simulations. An approximation of using chemical building blocks reduces the search phase space to manageable dimensions. This way the most favored structures can be located with as little computation as possible. Thus it is feasible to do structure search with large simulation cells, while still maintaining high chemical accuracy. The BOSS method was implemented as a python code called aalto-boss between 2016-2019, where I was the main author in co-operation with Milica Todorović and Patrick Rinke. I conducted a dimensional scaling study using analytic functions, which quantified the scaling of BOSS efficiency for fundamentally different functions when dimension increases. The results revealed the target function's derivative's important role to the optimization efficiency. The outcome will help people with choosing the simulation variables so that they are efficient to optimize, as well as help them estimate roughly how many BOSS iterations are potentially needed until convergence. The predictive efficiency and accuracy of BOSS was showcased in the conformer search of the alanine dipeptide molecule. The two most stable conformers and the characteristic 2D potential energy map was found with greatly reduced effort compared to alternative methods. The value of BOSS in novel materials research was showcased in the surface adsorption study of bifenyldicarboxylic acid on CoO thin film using DFT simulations. We found two adsorption configurations which had a lower energy than previous calculations and approximately supported the experimental data on the system. The three applications showed that BOSS can significantly reduce the computational load of atomistic structure search while maintaining predictive accuracy. It allows material scientists to study novel materials more efficiently, and thus help tailor the materials' properties to better suit the needs of modern devices.
  • Mäkelä, Noora (2022)
    Sum-product networks (SPN) are graphical models capable of handling large amount of multi- dimensional data. Unlike many other graphical models, SPNs are tractable if certain structural requirements are fulfilled; a model is called tractable if probabilistic inference can be performed in a polynomial time with respect to the size of the model. The learning of SPNs can be separated into two modes, parameter and structure learning. Many earlier approaches to SPN learning have treated the two modes as separate, but it has been found that by alternating between these two modes, good results can be achieved. One example of this kind of algorithm was presented by Trapp et al. in an article Bayesian Learning of Sum-Product Networks (NeurIPS, 2019). This thesis discusses SPNs and a Bayesian learning algorithm developed based on the earlier men- tioned algorithm, differing in some of the used methods. The algorithm by Trapp et al. uses Gibbs sampling in the parameter learning phase, whereas here Metropolis-Hasting MCMC is used. The algorithm developed for this thesis was used in two experiments, with a small and simple SPN and with a larger and more complex SPN. Also, the effect of the data set size and the complexity of the data was explored. The results were compared to the results got from running the original algorithm developed by Trapp et al. The results show that having more data in the learning phase makes the results more accurate as it is easier for the model to spot patterns from a larger set of data. It was also shown that the model was able to learn the parameters in the experiments if the data were simple enough, in other words, if the dimensions of the data contained only one distribution per dimension. In the case of more complex data, where there were multiple distributions per dimension, the struggle of the computation was seen from the results.
  • Koski, Jessica (2021)
    Acute lymphoblastic leukemia (ALL) is a hematological malignancy that is characterized by uncontrolled proliferation and blocked maturation of lymphoid progenitor cells. It is divided into B- and T-cell types both of which have multiple subtypes defined by different somatic genetic changes. Also, germline predisposition has been found to play an important role in multiple hematological malignancies and several germline variants that contribute to the ALL risk have already been identified in pediatric and familial settings. There are only few studies including adult ALL patients but thanks to the findings in acute myeloid leukemia, where they found the germline predisposition to consider also adult patients, there is now more interest in studying adult patients. The prognosis of adult ALL patients is much worse compared to pediatric patients and many are still lacking clear genetic markers for diagnosis. Thus, identifying genetic lesions affecting ALL development is important in order to improve treatments and prognosis. Germline studies can provide additional insight on the predisposition and development of ALL when there are no clear somatic biomarkers. Single nucleotide variants are usually of interest when identifying biomarkers from the genome, but also structural variants can be studied. Their coverage on the genome is higher than that of single nucleotide variants which makes them suitable candidates to explore association with prognosis. Copy number changes can be detected from next generation sequencing data although the detection specificity and sensitivity vary a lot between different software. Current approach is to identify the most likely regions with copy number change by using multiple tools and to later validate the findings experimentally. In this thesis the copy number changes in germline samples of 41 adult ALL patients were analyzed using ExomeDepth, CODEX2 and CNVkit.
  • Kähärä, Jaakko (2022)
    We study the properties of flat band states of bosons and their potential for all-optical switching. Flat bands are dispersionless energy bands found in certain lattice structures. The corresponding eigenstates, called flat band states, have the unique property of being localized to a small region of the lattice. High sensitivity of flat band lattices to the effects of interactions could make them suitable for fast, energy efficient switching. We use the Bose-Hubbard model and computational methods to study multi-boson systems by simulating the time-evolution of the particle states and computing the particle currents. As the systems were small, fewer than ten bosons, the results could be computed exactly. This was done by solving the eigenstates of the system Hamiltonian using exact diagonalization. We focus on a finite-length sawtooth lattice, first simulating weakly interacting bosons initially in a flat band state. Particle current is shown to typically increase linearly with interaction strength. However, fine-tuning the hopping amplitudes and boundary potentials, particle current through the lattice is highly suppressed. We use this property to construct a switch which is turned on by pumping the input with control photons. Inclusion of particle interactions disrupts the system, resulting in a large non-linear increase in particle current. We find that certain flat band lattices could be used as medium for an optical switch capable of controlling the transport of individual photons. In practice, highly optically nonlinear materials are required to reduce the switching time which is found to be inversely proportional to the interaction strength.
  • Koivula, Juho (2021)
    Kirjallisuuskatsauksessa käydään läpi erilaisia menetelmiä C3-substituoitujen indolien synteeseihin 2-alkenyylianiliinityyppisistä lähtöaineista, joiden bentsyylinen asema oli substituoitu. Erityistä huomiota kiinnitetään menetelmiin, joiden reaktiomekanismeiksi ehdotettiin radikaalimekanismeja. Myös näitä ehdotettuja radikaalimekanismeja esitellään tutkielmassa. Kokeellisessa työssä tutkittiin C3-subsituoitujen indolien hapettavaa synteesiä hiilikatalyytin avulla. Lähtöaineina käytettiin bentsyylisesti aryylisubstituoituja 2-alkenyylianiliinijohdannaisia. Joidenkin lähtöaineiden typpeen oli kiinnitetty metoksipyridiini, jonka kiinnitystä varten kehitettiin Buchwald-katalyysi. Hiilikatalyysit tuottivat hyviä saantoja. Korkea elektronitiheys, etenkin aniliinin bentseenirenkaan ja/tai typen aromaattisen substituentin korkea elektronitiheys, oli eduksi. Reaktion mekanismin ehdotetaan alkavan hapettumisella radikaalikationiksi, ja näitä hapetuspotentiaaleja laskettiin aiemmin raportoidun menetelmän mukaisesti. Mikäli indolin 5-renkaan substituutiot (N1, C2, C3) olivat tarpeeksi samankaltaisia, korkea elektronitiheys, matala hapetuspotentiaali ja hyvä saanto korreloivat. Indolin 5-renkaan substituutio on kuitenkin merkittävämpi tekijä kuin hapetuspotentiaali ja/tai korkea elektronitiheys. Pyridiini typen suojaryhmänä toimi katalyysissä ja se onnistuttiin poistamaan helposti. Metoksipyridiini toimi katalyysissä hyvin, mutta sen kvantitatiivinen poistaminen ei onnistunut.
  • Porna, Ilkka (2022)
    Despite development in many areas of machine learning in recent decades, still, changing data sources between the domain in a model is trained and the domain in the same model is used for predictions is a fundamental and common problem. In the area of domain adaptation, these circum- stances have been studied by incorporating causal knowledge about the information flow between features to be utilized in the feature selection for the model. That work has shown promising results to accomplish so-called invariant causal prediction, which means a prediction performance is immune to the change levels between domains. Within these approaches, recognizing the Markov blanket to the target variable has served as a principal workhorse to find the optimal starting point. In this thesis, we continue to investigate closely the property of invariant prediction performance within Markov blankets to target variable. Also, some scenarios with latent parents involved in the Markov blanket are included to understand the role of the related covariates around the latent parent effect to the invariant prediction properties. Before the experiments, we cover the concepts of Makov blankets, structural causal models, causal feature selection, covariate shift, and target shift. We also look into ways to measure bias between changing domains by introducing transfer bias and incomplete information bias, as these biases play an important role in the feature selection, often being a trade-off situation between these biases. In the experiments, simulated data sets are generated from structural causal models to conduct the testing scenarios with the changing conditions of interest. With different scenarios, we investigate changes in the features of Markov blankets between training and prediction domains. Some scenarios involve changes in latent covariates as well. As result, we show that parent features are generally steady predictors enabling invariant prediction. An exception is a changing target, which basically requires more information about the changes in other earlier domains to enable invariant prediction. Also, emerging with latent parents, it is important to have some real direct causes in the feature sets to achieve invariant prediction performance.
  • Tynkkynen, Jere (2022)
    This paper features two parts; a literature review discussing the recent development in using electrochemical gas sensors for pollutant detection and the use of sensor nodes in real-life locations, and an experimental section focusing on the kinetic study of nitrogen containing compounds utilizing in-tube extraction device. Growing interest towards personal safety have led to development of low-cost electrochemical sensors for personal safety, indoor air quality and leak detection applications. Heterojunctions and light illumination have emerged as an effective way to improve sensor performance, but the selectivity of electrochemical sensors remains relatively poor. Multiple sensors can be combined to create ‘E-noses’ which significantly improve the selectivity and compound identification. These E-noses have been deployed in some indoor locations, either being stationary in sensor networks or moved around by a robot or drone. All approaches have benefits and caveats associated to them, with the differences between individual sensors limiting sensor network use, and slow response and recovery times limiting the use of moving sensors. A novel micropump system was constructed to be used in the active air sampling together with in tube extraction (ITEX) and thermal desorption gas-chromatography (TD-GC-MS). The repeatability of this method was tested in a kinetic study of 10 selected nitrogen containing compounds in a custom-built permeation chamber. The breakthrough times and volumes of the compounds were investigated. Kinetic modelling was successful for 9 out of the 10 compounds with 1 compound behaving significantly different from the rest. The breakthrough times were always over 20 minutes and breakthrough volumes were around the 1000 ml region. Reproducibility was tested with multiple ITEX’s and samples were taken from five indoor locations. Three of the tested compounds were found in some of the samples.
  • Anttila, Jesse (2020)
    Visual simultaneous localization and mapping (visual SLAM) is a method for consistent self-contained localization using visual observations. Visual SLAM can produce very precise pose estimates without any specialized hardware, enabling applications such as AR navigation. The use of visual SLAM in very large areas and over long distances is not presently possible due to a number of significant scalability issues. In this thesis, these issues are discussed and solutions for them explored, culminating in a concept for a real-time city-scale visual SLAM system. A number of avenues for future work towards a practical implementation are also described.
  • Martinmäki, Tatu (2020)
    Tiivistelmä – Referat – Abstract Molecular imaging is visualization, characterization and quantification of biological processes at molecular and cellular levels of living organisms, achieved by molecular imaging probes and techniques such as radiotracer imaging, magnetic resonance imaging and ultrasound imaging. Molecular imaging is an important part of patient care. It allows detection and localization of disease at early stages, and it is also an important tool in drug discovery and development. Positron emission tomography (PET) is a biomedical imaging technique considered as one of the most important advances in biomedical sciences. PET is used for a variety of biomedical applications: i.e. imaging of divergent metabolism, oncology and neurology. PET is based on incorporation of positron emitting radionuclides to drug molecules. As prominent radionuclides used in PET are of short or ultra-short half-lives, the radionuclide is most often incorporated to the precursor in the last step of the synthesis. This has proven to be a challenge with novel targeted radiotracers, as the demand for high specific activity leads to harsh reaction conditions, often with extreme pH and heat which could denature the targeting vector. Click chemistry is a synthetic approach based on modular building blocks. The concept was originally developed for purposes of drug discovery and development. It has been widely utilized in radiopharmaceutical development for conjugating prosthetic groups or functional groups to precursor molecules. Click chemistry reactions are highly selective and fast due to thermodynamic driving force and occur with high kinetics in mild reaction conditions, which makes the concept ideal for development and production of PET radiopharmaceuticals. Isotope exchange (IE) radiosynthesis with trifluoroborate moieties is an alternative labeling strategy for a reasonably high yield 18F labeling of targeted radiopharmaceuticals. As the labeling conditions in IE are milder than in commonly utilized nucleophilic fluorination, the scope of targeting vectors can be extended to labile biomolecules expressing highly specific binding to drug targets, resulting to higher contrast in PET imaging. A trifluoroborate functionalized prosthetic group 3 was synthetized utilizing click chemistry reactions, purified with SPE and characterized with HPLC-MS and NMR (1H , 11B-, 13C-, 19F-NMR). [18F]3 was successfully radiolabeled with RCY of 20.1 %, incorporation yield of 22.3 ± 11.4 % and RCP of >95 %. TCO-functionalized TOC-peptide precursor 6 was synthetized from a commercial octreotide precursor and a commercially available click chemistry building block via oxime bond formation. 6 was characterized with HPLC-MS and purified with semi preparative HPLC. Final product [18F]7 was produced in a two-step radiosynthesis via IEDDA conjugation of [18F]3 and 6. [18F]7 was produced with RCY 1.0 ± 1.0 %, RCP >95 % and estimated molar activity of 0.7 ± 0.8 GBq/µmol. A cell uptake study was conducted with [18F]7 in AR42J cell line. Internalization and specific binding to SSTR2 were observed in vitro.
  • Laaksonen, Jenniina (2021)
    Understanding customer behavior is one of the key elements in any thriving business. Dividing customers into different groups based on their distinct characteristics can help significantly when designing the service. Understanding the unique needs of customer groups is also the basis for modern marketing. The aim of this study is to explore what types of customer groups exist in an entertainment service business. In this study, customer segmentation is conducted with k-prototypes, a variation of k-means clustering. K-prototypes is a machine learning approach partitioning a group of observations into subgroups. These subgroups have little variation within the group and clear differences when compared to other subgroups. The advantage of k-prototypes is that it can process both categorical and numeric data efficiently. The results show that there are significant and meaningful differences between customer groups emerging from k-prototypes clustering. These customer groups can be targeted based on their unique characteristics and their reactions to different types of marketing actions vary. The unique characteristics of the customer groups can be utilized to target marketing actions better. Other possibilities to benefit from customer segmentation include such as personalized views, recommendations and helping strategy level decision making when designing the service. Many of these require further technical development or deeper understanding of the segments. Data selection as well as the quality of the data has an impact on the results and those should be considered carefully when deciding future actions on customer segmentation.
  • Koivisto, Teemu (2021)
    Programming courses often receive large quantities of program code submissions to exercises which, due to their large number, are graded and students provided feedback automatically. Teachers might never review these submissions therefore losing a valuable source of insight into student programming patterns. This thesis researches how these submissions could be reviewed efficiently using a software system, and a prototype, CodeClusters, was developed as an additional contribution of this thesis. CodeClusters' design goals are to allow the exploration of the submissions and specifically finding higher-level patterns that could be used to provide feedback to students. Its main features are full-text search and n-grams similarity detection model that can be used to cluster the submissions. Design science research is applied to evaluate CodeClusters' design and to guide the next iteration of the artifact and qualitative analysis, namely thematic synthesis, to evaluate the problem context as well as the ideas of using software for reviewing and providing clustered feedback. The used study method was interviews conducted with teachers who had experience teaching programming courses. Teachers were intrigued by the ability to review submitted student code and to provide more tailored feedback to students. The system, while still a prototype, is considered worthwhile to experiment on programming courses. A tool for analyzing and exploring submissions seems important to enable teachers to better understand how students have solved the exercises. Providing additional feedback can be beneficial to students, yet the feedback should be valuable and the students incentivized to read it.
  • Pritom Kumar, Das (2024)
    Many researchers use dwell time, a measure of engagement with information, as a universal measure of importance in judging the relevance and usefulness of information. However, it may not fully account for individual cognitive variations. This study investigates how individual differences in cognitive abilities, specifically working memory and processing speed, can significantly impact dwell time. We examined the browsing behavior of 20 individuals engaged in information-intensive tasks, measuring their cognitive abilities, tracking their web page visits, and asses their perceived relevance and usefulness of the information encountered. Our findings shows a clear connection between working memory, processing speed, and dwell time. Based on this finding, we developed a model that combines both cognitive abilities and dwell time to predict the relevance and usefulness of web pages. Interestingly, our analysis reveals that cognitive abilities, particularly fluid intelligence, significantly influence dwell time. This highlights the importance of incorporating individual cognitive differences into prediction models to improve their accuracy. Thus, personalized services that set dwell time thresholds based on individual users' cognitive abilities could provide more accurate estimations of what users find relevant and useful.
  • Brasseur, Paul (2021)
    Plasmonic is an emerging field which has showed application for photocatlysis. Here we investigate a gold/platinum bimetallic catalytic system, and try to show how the catalytic properties of gold nanoparticles can be us to harvest visible light energy to increase the catalytic activity of platinum. Platinum being are rare and expensive metal, we also took the opportunity to find the optimal amount of catalyst to reduce platinum use. The catalyst is composed of a core spherical gold nanoparticles, of around 15 nm diameter. They were synthesized using an inversed Turkevich method, based on trisodium citrate, gold precursor salt and done in solution. Various amount of platinum was deposited on those nanoparticles using seeded growth method. The amount of platinum varied for single atoms to an atomic monolayer. This suspension of nanoparticles was deposited on ultrafine silica powder to be used for certain reaction and characterization. The material was characterized via several technics. UV-Visible and Diffuse Reflectance Spectroscopy were used to characterize its optical properties and showed a absorption peak around 524 nm characteristic of gold nanoparticles of this size. Imaging was done using electron microscopy (SEM and TEM) to study the morphology and showed monodisperse and spherical particles. The exact composition of the different catalyst were obtain using Atomic Emission Spectroscopy. The study was conducted by using reduction reaction as tests to investigate differences in conversion and selectivity under dark and monochromatic 525 nm and 427 nm light conditions. We chose to work on reduction of 4-nitrophenol, phenylacetylene and nitrobenzene, because they are widely used both in research and industry, and are easy to set up. Some catalyst showed good enhancement under 525 nm light, especially the one with the least amount of platinum. Different selectivity were also observed, indicating the presence of different reaction pathways under light conditions.
  • Hyvärinen, Linda (2023)
    With the increased usage of machine learning models in various tasks and domains, the demand of understanding the models is emphasized. However, often modern machine learning models are difficult to understand and therefore do not provoke trust. Models can be understood by revealing their inner logic with explanations, but explanations can be difficult to interpret for non-expert users. We introduce an interactive visual interface to help non-expert users to understand and compare machine learning models. The interface visualizes explanations for multiple models in order to help the user to understand how the models generate predictions and whether the predictions can be trusted. We also explore current research in explainable AI visualizations, in order to compare our prototype to comparable systems present in research. The contributions of this paper are a system description and a use case for an interactive visualization interface to compare and explain machine learning models, as well as providing an understanding of the current state of research in explainable AI visualization systems and recommendations for future studies. We conclude that our system enables efficient visualizations for regression models unlike the papers covered in our survey. Another conclusion is that the field lacks precise terminology.
  • Jääskeläinen, Matias (2020)
    This thesis is about exploring descriptors for atmospheric molecular clusters. Descriptors are needed for applying machine learning methods for molecular systems. There is a collection of descriptors readily available in the DScribe-library developed in Aalto University for custom machine learning applications. The question of which descriptors to use is up to the user to decide. This study takes the first steps in integrating machine learning into existing procedure of configurational sampling that aims to find the optimal structure for any given molecular cluster of interest. The structure selection step forms a bottleneck in the configurational sampling procedure. A new structure selection method presented in this study uses k-means clustering to find structures that are similar to each other. The clustering results can be used to discard redundant structures more effectively than before which leaves fewer structures to be calculated with more expensive computations. Altogether that speeds up the configurational sampling procedure. To aid the selection of suitable descriptor for this application, a comparison of four descriptors available in DScribe is made. A procedure for structure selection by representing atmospheric clusters with descriptors and labeling them into groups with k-means was implemented. The performance of descriptors was compared with a custom score suitable for this application, and it was found that MBTR outperforms the other descriptors. This structure selection method will be utilized in the existing configurational sampling procedure for atmospheric molecular clusters but it is not restricted to that application.
  • Nissilä, Viivi (2020)
    Origin-Destination (OD) data is a crucial part of price estimation in the aviation industry, and an OD flight is any number of flights a passenger takes in a single journey. OD data is a complex set of data that is both flow and multidimensional type of data. In this work, the focus is to design interactive visualization techniques to support user exploration of OD data. The thesis work aims to find which of the two menu designs suit better for OD data visualization: breadth-first or depth-first menu design. The two menus follow Schneiderman’s Task by Data Taxonomy, a broader version of the Information Seeking Mantra. The first menu design is a parallel, breadth-first menu layout. The layout shows the variables in an open layout and is closer to the original data matrix. The second menu design is a hierarchical, depth-first layout. This layout is derived from the semantics of the data and is more compact in terms of screen space. The two menu designs are compared in an online survey study conducted with the potential end users. The results of the online survey study are inconclusive, and therefore are complemented with an expert review. Both the survey study and expert review show that the Sankey graph is a good visualization type for this work, but the interaction of the two menu designs requires further improvements. Both of the menu designs received positive and negative feedback in the expert review. For future work, a solution that combines the positives of the two designs could be considered. ACM Computing Classification System (CCS): Human-Centered Computing → Visualization → Empirical Studies in Visualization Human-centered computing → Interaction design → Interaction design process and methods → Interface design prototyping
  • Rahikainen, Tintti (2023)
    Machine learning operations (MLOps) tools and practices help us continuously develop and de- ploy machine learning models as part of larger software systems. Explainable machine learning can support MLOps, and vice versa. The results of machine learning models are dependent on the data and features the models use, so understanding the features is important when we want to explain the decisions of the model. In this thesis, we aim to understand how feature stores can be used to help understand the features used by machine learning models. We compared two existing open source feature stores, Feast and Hopsworks, from an explainability point of view to explore how they can be used for explainable machine learning. We were able to use both Feast and Hopsworks to aid us in understanding the features we extracted from two different datasets. The feature stores have significant differences, Hopsworks being a part of a larger MLOps platform, and having more extensive functionalities. Feature stores provide useful tools for discovering and understanding the features for machine learning models. Hopsworks can help us understand the whole lineage of the data – where it comes from and how it has been transformed – while Feast focuses on serving the features consistently to models and needs complementing services to be as useful from an explainability point of view.