Skip to main content
Login | Suomeksi | På svenska | In English

Browsing by Title

Sort by: Order: Results:

  • Karvo, Sara (2023)
    Zooplankton are an important link in marine pelagic food webs as they transfer energy from primary producers to higher trophic levels such as planktivorous fish. They migrate vertically in the water column, ascending to feed near the surface at night and descending to hide from visual predators for the day (diel vertical migration, DVM). Zooplankton are detected with Acoustic Doppler Current Profilers (ADCPs). These devices were developed for measuring water currents using acoustic pulses, a technique which requires particles such as zooplankton in the water column to scatter the sound. As a by-product of the velocity measurements, it provides information of these scatterers as echo intensity. This method has been used in researching zooplankton DVM, however, not in the northern Baltic Sea prior to this study. In this thesis, the data processing steps required to analyze echo intensity were examined for the specific environment of the Finnish Archipelago Sea. A one-year-long time-series was processed and averaged seasonally to investigate different patterns in zooplankton DVM. Vertical velocity data were used in estimating migration speed, and available reference measurements were combined to the data to examine the environmental factors affecting zooplankton DVM. Synchronized DVM was observed especially in autumn, however, indications of other migration patterns such as unsynchronized and reverse migration were detected during summer and winter, respectively. The primary cue behind zooplankton DVM was light, but additional contributing factors such as phytoplankton and currents were identified and discussed. The maximum migration speeds detected were approximately 10 cm/s downwards and 4 cm/s upwards. ADCP data are a good indicator of zooplankton migration in the northern Baltic Sea and in the future, it could prove beneficial in zooplankton monitoring and biomass estimates.
  • Bagchi, Rupsha (2017)
    The Internet of Things is a proliferating industry, which is transforming many homes and businesses, making them smart. However, the rapid growth of these devices and the interactions between these devices, introduces many challenges including that of a secure management system for the identities and interactions of the devices. While the centralized model has worked well for many years, there is a risk of the servers becoming bottlenecks and a single point of failure, thereby making them vulnerable to Denial-of-Service attacks. As a backbone of these interactions, Blockchain is capable of creating a highly secure, independent and distributed platform. Blockchain is a peer to peer, distributed ledger system that stores all the transactions taking place within the network. The main purpose of the servers that form a part of the distributed system is to provide a consensus, using various consensus algorithms, on the state of the blockchain at any given time and to store a copy of all the transactions taking place. This thesis explores the Blockchain technology in general and investigates its potential with regard to access management of constrained devices. A proof of concept system has been designed and implemented that demonstrates a simplified access management system using Ethereum Blockchain. This was done to check whether the concept can be applied at a global level. Although the latency of the network depends on the computing power of the resources participating in the Blockchain, an evaluation of the proof of concept system has been made, keeping in mind the smallest device that can be involved in the consensus process. Docker containers have been used to simulate a cluster of the nodes participating in the Blockchain, in order to examine the implemented system. An outline of the various advantages and the limitations of Blockchains in general, as well as the developed proof of concept system, has also been provided.
  • Leppänen, Leo (2017)
    In this study we use element-level usage data that was collected from the online learning material of an university level introductory programming course for identification of areas-of-interest in the course material as well as for prediction of student learning outcomes. The data was collected in-situ using a JavaScript component embedded in the online learning material, which recorded which HTML elements were visible on the user's screen after each interaction (movement and click) and if the user's screen had been still for at least 2500 milliseconds. A visual analysis indicates that students spend large amounts of time on material sections that discuss special syntactic structures that they are unable to infer from previous experience. Overall, the analysis was able to identify areas of the online learning material that seem to be too long and in-depth for the concepts they are discussing, when the things the students have previously learned are taken into account. This high-level analysis also revealed that the time the students spent viewing an assignment's prompt was statistically significantly correlated with the perceived workload, difficulty and educational value of that same assignment. We observe that when partial correlations are considered, and multiple comparisons are corrected for, time spent with an assignment's prompt on the screen is no longer statistically significantly correlated with the three variables. The same usage data was used to investigate whether material usage statistics can predict learning outcomes or identify strong and at-risk students. The results indicate that based on just three to four weeks of data, it is possible to identify strong and at-risk students with some accuracy. Furthermore, it seems possible to identify student programming assignment scores and total course scores with a somewhat high accuracy. Models based on material usage statistics also displayed some light predictive power in predicting student exam scores. It was also shown that the predictive powers of these models are not based solely on student effort or time-on-task. All told, this thesis demonstrates that fine-grained online learning material usage data is feasible to collect and useful in understanding both the students and the learning material. The results suggest that very simple and almost entirely domain-independent data sources can be used to predict student performance to a relatively large degree, suggesting that a combination of such simple domain-independent metrics could match highly domain dependent and more complex metrics in predictive power, giving raise to more widely usable educational analytics tools.
  • Äijälä, Cecilia (2020)
    Tropes are storytelling devices or conventions that can be found in storytelling media, for example in movies. DBTropes is an RDF-dataset with media-trope connections extracted from Tv Tropes, which is a community-edited wiki listing tropes for various creative works. This study investigates whether the tropes of films can be used to cluster similar movies. First, we extracted film-trope connections from the DBTropes dataset. We then took four samples from the dataset, three for clustering films, and one for clustering tropes. We used the film–trope connections to calculate euclidean and cosine distances between movies and for the last sample between tropes. Then we clustered the samples with hierarchical clustering using complete and ward linkage. Our results show that hierarchical clustering can group similar films together using this dataset. For calculating distances the cosine distance method works significantly better than euclidean distance. Both hierarchical clustering methods complete and ward work well. It depends on the chosen sample, which of them results in a clearer and more interpretable output. We conclude that the data works well for clustering similar films or similar tropes.
  • Hästbacka, Matti (2023)
    The direct economic impacts of the global tourism industry account for 4 % of global GDP and 8 % of global greenhouse gas emissions. The industry is in transformation caused by climate change, political instability and rapid technological development. In addition, the relationship between biodiversity conservation and tourism as well as the growing popularity are considered megatrends impacting the sector. Traditional mass tourism destinations, such as the Canary Islands, may start seeing new kinds of visitors, if traveling to exotic destinations becomes difficult as a result of these transformations. Understanding transformations affecting tourism requires information about tourists’ mobilities, interests and preferences. However, traditional data collection methods may not necessarily be suited for studying quickly changing tourism. The need for Information about visitations to natural and protected areas is especially high, as traditional tourism indicators, such as flights and accommodation statistics do not tell where the tourists spend time. Social media data may enable production of new kind of knowledge and studying nature-based tourism in a new way. In this thesis, I intent to assess the role of nature in tourism in the Canary Islands, Spain using data from the photo-sharing platform Flickr. First, I compare the spatiotemporal patterns of Flickr data against official data about tourism flows to confirm the feasibility of Flickr as a data source in the Canary Islands context. I then try to understand the importance of nature visitations and differences in nature visitation patterns between visitors from different countries. Finally, I turn to analyse contents of the images to see what kinds of nature-related topics are important for each group, making use of a deep learning and cluster detection algorithms. I verify the results of my empirical analysis with data collected through interviewing experts familiar with Canary Islands tourism. The results of my research show that Flickr reflects Canary Islands tourism patterns moderately well, and that it can be used to produce information about differences in nature visitation patterns. Protected areas are shown to be important and central for Canary Islands tourism, but differences in interest toward these areas between groups are notable. Results of the content analyses show that while differences between groups exist, both nature-related content and photos of humans are important in content posted from PAs. Verification data collected through expert interviews shows that the observed differences between groups correspond to the experts’ perceptions about differences between different groups. The findings of my thesis demonstrate the importance of nature and protected areas in Canary Islands tourism and confirm earlier knowledge about the use of Flickr in studying nature visitations. The results may inform future research in the Canary Islands. More broadly, they provide information about the feasibility and limitations of the use of social media data for nature-based tourism research.
  • Ba, Yue (2021)
    Ringed seals (Pusa hispida) and grey seals (Halichoerus grypus) are known to have hybridized in captivity despite belonging to different taxonomic genera. Earlier genetic analyses have indicated hybridization in the wild and the resulting introgression of genetic material cross species boundaries could potentially explain the intermediate phenotypes observed e.g. in their dentition. Introgression can be detected using genome data, but existing inference methods typically require phased genotype data or cannot separate heterozygous and homozygous introgression tracts. In my thesis, I will present a method based on Hidden Markov Models (HMM) to identify genomic regions with a high density of single nucleotide variants (SNVs) of foreign ancestry. Unlike other methods, my method can use unphased genotype data and can separate heterozygous and homozygous introgression tracts. I will apply this method to study introgression in Baltic ringed seals and grey seals. I will compare our method to an alternative method and assess our method with simulated data in terms of precision and recall. Then, I will apply it to seal data to search for introgression. Finally, I will discuss what future directions to improve our method.
  • Laitinen, Emma (2023)
    Implementing software process improvement (SPI) models or standards can be challenging for a small organization due to their limited resources compared to larger companies. The ISO/IEC 29110 series of systems and engineering standards were designed especially for very small entities (VSEs), i.e. for organizations having up to 25 employees. Company X is a small Finnish software company following a Scrum workflow. At Company X, challenges have been identified in the software testing process. Because of the company’s size, ISO/IEC 29110 out of different SPI standards was identified as a potential fit for improving this process. While the ISO/IEC 29110 standard can be applied to any software life cycle method, including agile, there is no formal guide on how to implement the standard in an agile environment. The aims of this thesis are two-fold: first, to investigate how Scrum corresponds with the standard, and second, to use the standard to identify weak points in Company X’s current software testing process and to identify action points to address them. The mappings between Scrum and the standard were investigated by carrying out a systematic literature review (SLR). A self-assessment and a software testing deployment package provided with the standard were used to assess the current testing process and to identify shortcomings in it. The shortcomings were analyzed and action points feasible in Company X’s context were suggested. The improved process containing the suggested action points was then re-assessed. The SLR yielded only a handful of papers, indicating that the topic of implementing ISO/IEC 29110 into an agile lifecycle in practice is relatively unexplored. The three papers together provided mappings for all three aspects of the standard vs. their counterparts in Scrum: activities, roles, and work products. The baseline assessment of Company X’s current process yielded a score of achieved ‘Partially’ (46,5 %). A set of seven shortcomings were identified in the assessment process and nine action points were suggested to address them. Assessing the improved process improved the score to implemented ‘Fully’ (97 %).
  • Niiranen, Juha (2016)
    The demand for mobile services is increasing constantly and mobile network operators need to significantly upgrade their networks to respond to the demand. The increasing complexity of the networks makes it impossible for a human operator to manage them optimally. Currently the network management operations are automated using a pre-defined logic. The future goal is to introduce cognitive network management functions which can adapt to changes in the network context and handle uncertainty in network data. This thesis discusses using Markov Logic Networks for cognitive management of mobile networks. The method allows uncertain and partial information and makes it possible to consolidate knowledge from multiple sources into a single, compact, representation. The model can be used to infer configuration changes in network parameters and the model parameters can be learned from data. We test the method in a simulated LTE network and examine the results in terms of improvements in network performance and computational cost.
  • Karvonen, Mikko (Helsingin yliopistoHelsingfors universitetUniversity of Helsinki, 2008)
    The usual task in music information retrieval (MIR) is to find occurrences of a monophonic query pattern within a music database, which can contain both monophonic and polyphonic content. The so-called query-by-humming systems are a famous instance of content-based MIR. In such a system, the user's hummed query is converted into symbolic form to perform search operations in a similarly encoded database. The symbolic representation (e.g., textual, MIDI or vector data) is typically a quantized and simplified version of the sampled audio data, yielding to faster search algorithms and space requirements that can be met in real-life situations. In this thesis, we investigate geometric approaches to MIR. We first study some musicological properties often needed in MIR algorithms, and then give a literature review on traditional (e.g., string-matching-based) MIR algorithms and novel techniques based on geometry. We also introduce some concepts from digital image processing, namely the mathematical morphology, which we will use to develop and implement four algorithms for geometric music retrieval. The symbolic representation in the case of our algorithms is a binary 2-D image. We use various morphological pre- and post-processing operations on the query and the database images to perform template matching / pattern recognition for the images. The algorithms are basically extensions to classic image correlation and hit-or-miss transformation techniques used widely in template matching applications. They aim to be a future extension to the retrieval engine of C-BRAHMS, which is a research project of the Department of Computer Science at University of Helsinki.
  • Vartiainen, Panu (2014)
    The thesis discusses possibilities for using metadata and context information in annotating, sharing, and searching user-created content in the mobile domain. The first part of the thesis discusses metadata, ontologies, context information, and imaging. The latter part of the thesis describes a prototype system for classifying and annotating digital photographs and storing context information as metadata of the photographs in a mobile phone. Another role of the prototype system is to perform context- and ontology-based information retrieval using a mobile phone user interface. The prototype system contains a limited RDF metadata engine and an ontology browser for mobile phones, as well as a server-side metadata and content repository. The implementation demonstrates that a part of the creation-time context, such as the location and temporal context, can be automatically gathered in a mobile phone, and stored as metadata for the content. In addition, the same parts of context information can be used for searching. The content and the metadata can be stored on a server and shared with other users. The prototype is built around a tourism scenario that works as an example of how these technologies can be used in a mobile phone.
  • Hussain, Zafar (2020)
    The National Library of Finland has digitized newspapers starting from late eighteenth century. Digitized data of Finnish newspapers is a heterogeneous data set, which contains the content and metadata of historical newspapers. This research work is focused to study this rich materiality data to find the data-driven categorization of newspapers. Since the data is not known beforehand, the objective is to understand the development of newspapers and use statistical methods to analyze the fluctuations in the attributes of this metadata. An important aspect of this research work is to study the computational and statistical methods which can better express the complexity of Finnish historical newspaper metadata. Exploratory analyses are performed to get an understanding of the attributes and extract the patterns among them. To explicate the attributes’ dependencies on each other, Ordinary Least Squares and Linear Regression methods are applied. The results of these regression methods confirm the significant correlation between the attributes. To categorize the data, spectral and hierarchical clustering methods are studied for grouping the newspapers with similar attributes. The clustered data further helps in dividing and understanding the data over time and place. Decision trees are constructed to split the newspapers after attributes’ logical divisions. The results of Random Forest decision trees show the paths of development of the attributes. The goal of applying various methods is to get a comprehensive interpretation of the attributes’ development based on language, time, and place and evaluate the usefulness of these methods on the newspaper data. From the features’ perspective, area appears as the most imperative feature and from language based comparison Swedish newspapers are ahead of Finnish newspapers in adapting popular trends of the time. Dividing the newspaper publishing places into regions, small towns show more fluctuations in publishing trends, while from the perspective of time the second half of twentieth century has seen a large increase in newspapers and publishing trends. This research work coordinates information on regions, language, page size, density, and area of newspapers and offers robust statistical analysis of newspapers published in Finland.
  • Rychkova, Kseniya (2022)
    The Traveling Salesman Problem (TSP) is a well-known optimization problem. The time needed to solve TSP classically grows exponentially with the size of the input, placing it into the NP-hard computational complexity class–the class of problems that are at least as hard as any other problem solvable in nondeterministic polynomial time. Quantum computing gives us a new approach to searching through such a huge search space, using methods such as quantum annealing and phase estimation. Although the current state of quantum computers does not give us enough resources to solve TSP with a large input, we can use quantum computing methods to improve on existing classical algorithms. The thesis reviews existing methods to efficiently tackle TSP utilizing potential quantum resources, and discusses the augmentation of classical algorithms with quantum techniques to reduce the time complexity of solving this computationally challenging problem.
  • Niinikoski, Eerik (2020)
    The aim of this thesis is to predict total career racing performance of Finnish trotter horses by using trotters early career racing performance and other early career variables. This thesis presents a brief introductory of harness racing and horses used in Finnish trotting sport. The data is presented and modified for predictions, with descriptive statistics of tables and visuals. The machine learning method of Random forests for regression is introduced and used in the predictions. After training the model, this thesis presents the prediction accuracy and variables of importance of the predictions of total career racing performance for both Finnhorse trotters and Finnish Standardbred trotter population. Finally, the writer discusses on the shortages and possible improvements for future research. The data for this thesis was provided by The Finnish trotting and breeding association (Suomen Hippos ry), which included all information of harness races from 1984 to the end of 2019, raced in Finland. From almost three million rows, the data was summarised to a data table of 46704 rows of trotters, that have started their career at earliest allowed three age groups. A total of 37 independent variables were used to predict three outcomes of total career earnings, total number of career starts and total number of career first placings, as separate models. The predictors are derived from other studies that estimate the environmental and genetic factors of racing performance of a trotter. The three models performed poor to moderate, with total earnings having the highest prediction accuracy. The model predicted quite well larger amounts of earnings, but was avid to predict some earnings when there in fact were none. Prediction accuracy of total number of starts was poor, especially when the true amount of starts was low. Model that predicted total number of career first placings performed the worst. This can partially be explained by the fact that winning is a rare event for a trotter in general. The models fit better for Finnish Standardbred trotters than for Finnhorse trotters. This thesis works as a good basis for future similar research, where massive amounts of data and machine learning is used to predict trotter’s career, racing performance or other factors. The results show that predicting total career racing performance as a classification problem could be a better fit than regression. These adequate classes, as well as possible better predictors and suitable imputes for missing values, should be consulted with an audience of superior knowledge in harness racing.
  • Kropotov, Ivan (2020)
    Reinforcement learning (RL) is a basic machine learning method, which has recently gained in popularity. As the field matures, RL methods are being applied on progressively more complex problems. This leads to need to design increasingly more complicated models, which are difficult to train and apply in practice. This thesis explores one potential way of solving the problem with large and slow RL models, which is using a modular approach to build the models. The idea behind this approach is to decompose the main task into smaller subtasks and have separate modules each of which concentrates on solving a single subtask. In more detail, the proposed agent will be built using the Q-decomposition algorithm, which provides a simple and robust algorithm for building modular RL agents. The problem we use as an example of usefulness of the modular approach is a simplified version of the video game Doom and we design a RL agent that learns to play it. The empirical results indicate that the proposed model is able to learn to play the simplified version of Doom on a reasonable level, but not perfectly. Additionally, we show that the proposed model might suffer from usage of too simple models for solving the subtasks. Nevertheless, taken as a whole the results and the experience of designing the agent show that the modular approach for RL is a promising way forward and warrants further exploration.
  • Lobo, Hannah (2021)
    The lidar depolarisation ratio is used for aerosol categorisation as it is indicative of aerosol shape. Commonly, depolarisation ratio is measured in short term studies at short wavelengths such as 355 nm and 532 nm. The depolarisation ratio has a spectral dependency and so exploring values at longer wavelengths could be valuable for future studies. Here, aerosol depolarisation ratio at 1565 nm is measured across Finland’s ground based remote sensing network over a four year period. The Halo Photonics StreamLine Doppler lidars instruments were found to be stable over long time periods and cloud based calibration was used to correct for the bleed though. The depolarisation ratio of elevated aerosol layers was compared to boundary layer aerosol. A higher average depolarisation ratio was found for elevated aerosol with the exception of boreal forest sites in the summer months where values were similar. Elevated aerosols over Finland were found to originate mostly from the Arctic, Europe, Russia and North America using aerosol transport models. Four case studies were looked at in more detail: Saharan dust with a depolarisation ratio of 0.249 ± 0.018, pollen with a depolarisation ratio of 0.207 ± 0.013, anthropogenic pollution with a depolarisation ratio of 0.067 ± 0.009, and a mixed layer with a depolarisation ratio of 0.152 ± 0.019 thought to be pollen and smoke. Based on this study, Halo Doppler Lidar can be used to measure elevated aerosol at 1565 nm in the long term. Future studies could use 1565 nm depolarisation ratio alongside commonly used shorter wavelengths to aid aerosol categorisation.
  • Jebaraj, Immanuel Christopher (2017)
    Distinguishing the coronal magnetic field and its evolution can unlock key information on solar energetic eruptions such as the Coronal Mass Ejections (CMEs). CMEs are formed as magnetic flux ropes, i.e. magnetic field lines twisted about each other. They are the main drivers of space weather effects on Earth. Understanding the structure of the internal magnetic field of the CME would help determine the severity of the resulting geomagnetic storm. Predicting the onset and the orientation of the flux rope axis is a major focus of current space weather research. For this purpose, a numerical study on the kinematic emergence of a twisted flux rope into a coronal magnetic field is performed using the Magneto-frictional method (MFM). The MFM is an exciting prospect as it is sufficiently accurate and computationally inexpensive. The initiation of the eruption is through ideal Magnetohydrodynamic (MHD) kink instability. In this case, the kink instability occurs when the windings of the field lines about the flux rope axis exceeds a critical value. This thesis presents the set-up of the Fan & Gibson flux rope with different configurations. This was in hopes of studying the slow energization of the coronal field arcade with the emergence of a current carrying flux rope. The results of the simulations presented here show that the several key factors such as the height at which the flux rope is stopped and its twist play a major role in the dynamics of the flux rope in making it kink unstable. One of the main motivations was to use the results to discuss the performance of the MFM in comparison to MHD and how capable it is in capturing ideal MHD phenomenon. The simulations are also used to investigate the formation of sigmoidal current layer often seen before the onset of eruption. In the results presented here, the sigmoidal ’S’ shaped current layer is formed as the flux rope becomes kink unstable. This sigmoidal current layer is analysed for different configurations of the flux rope. These results have suggested that accurate dynamic modelling of the coronal magnetic field is essential for successful space weather prediction purposes.
  • Väkevä, Sakari (2019)
    The reflection seismic surveying method is useful when conducting mineral exploration in the crystalline bedrock because of its good depth extent and resolution. However, the traditional experiments with active sources are expensive and difficult to carry out, especially in remote areas or in conservation areas where mineral exploration is limited due to environmental reasons. Recently, a number of theoretical advances have proven that passive soundings utilizing ambient seismic noise can provide new opportunities for seismic imaging and contribute to data generation for reflection seismic surveys, without the need for explosive or vibratory sources. One of the most promising new methods is seismic interferometry (SI), where the impulse response between two receivers is reconstructed by correlating their signals with each other. COGITO-MIN is a joint project between the University of Helsinki, the Geological Survey of Finland, Polish Academy of Sciences, and industrial partners with the aim of investigating and developing new cost-effective seismic exploration methods in the crystalline bedrock. Within the framework of the project, a passive seismic experiment was carried out in which 45 three-component geophones were deployed for a month in the vicinity of the polymetallic Kylylahti Mine in Polvijärvi, northern Karelia, where the mining operator is the Swedish metal company Boliden. The original purpose of these geophones was to collect data suitable for detecting underground cavities related to underground nuclear explosions. The institute that collected the data was CTBTO (Comprehensive Test Ban Treaty Organization) whose task is to monitor the treaty in the pre-ratification stage. The purpose of this Master's thesis was to develop an SI workflow for the three-component data and to investigate the method's performance in an area where local geology is known after nearly 40 years of exploration and consequent mining operations. The specific scientific objectives of the thesis are (1) to demonstrate the usefulness of collecting three-component data in conjunction with or instead of single-component data, (2) to assess the noise-based SI methods used in previous studies and to improve their stability in the crystalline bedrock, and (3) to investigate the possibilities of SI from an operational perspective. Seismic velocities obtained through laboratory measurements were merged with geological and density models of the target area provided by Boliden. The resulting velocity and density grids were then used as the basis for waveform modelling, and the results from SI were validated against them. The starting point for SI was the noise-driven approach where 'each sample matters'. The interferometric workflow is built on the Seismic Unix suite together with self-written algorithms that are based on theoretical evaluations. SI is followed by an imaging workflow, which provides the basis for the reflectivity profiles. The thesis work focuses on five components of the Green's tensor and the vertical, radial and transverse component of the impulse response. With the horizontal components, one can access the S-wave patterns in addition to the P-waves. As a specialty, the so-called sign bit normalization (SBN) method was also tested. The technique involves destroying much of the amplitude information of the original seismograms by only retaining the sign bit of each sample. According to the results outlined in this thesis, SBN can make it easier to image the weak reflectors of the subsurface. This type of seismic interferometry seems particularly suitable for the early stage of mineral exploration, where the explorer does not yet fully understand the target they are studying. The most important advantage of seismic interferometry, however, is its cost effectiveness, and its potential for reducing risks for the environment.
  • Westlin, Emilia (2022)
    The aim of this thesis was to 1) give an exposition of how topological data analysis (TDA) can be used to look for patterns in periodic data, 2) apply it to financial data and 3) visually explore how a topological analysis of credit data using landscape distances compared to looking directly at the change in credit data in the context of stock market crashes. TDA applies algebraic topology to data. It models data sets as various-dimensional surfaces, or manifolds, and studies their structure to find patterns of interconnectedness. It is a powerful tool for studying large, complex, multi-dimensional and noisy data sets. It is often able to capture subtle patterns in such data sets much better than other methods. It is known that stock market crashes are preceded by periods of credit expansion, but we have no reliable indicator of an imminent crash. Chapter 2 covers the algebraic topological theory needed. Key concepts are simplicial complexes, homology groups and persistent homology. The central theorem is the Nerve Theorem, which establishes an equivalence between the union of a collection of convex sets and the nerve of the collection. Chapter 3 describes the method of time delay embedding to pre-process periodic data. A Vietoris-Rips filtration was applied to sliding windows of credit data. From this persistence diagrams and their corresponding persistence landscapes were obtained. The normalised persistence landscape norms (L1) were plotted to visually explore how well TDA captured the connection between credit expansion and stock market crashes. It was compared to the discrete first derivative of the credit data. Visual inspection of the graphs suggested TDA to be as good, and possibly slightly better, at predicting stock market crashes from bank credit data, than looking at the discrete first derivative directly. No obvious new indicator of an imminent crash was found, however. To unlock the true potential of TDA in analysing large, multivariate data sets, further studies could look to triangulate a better indicator of stock market crashes by combining the credit data with other economic, social and political data. It would also be useful to establish a less subjective, more transparent method for choosing the thresholds used as crash indicators, and to quantify the predictions made by different indicators to better compare them with each other.
  • Joutsenvirta, Timo (2020)
    Ohjelmiston hyvä käytettävyys varmistaa sen, että henkilöt, joille ohjelmisto on suunnattu osaavat sitä käyttää, ja että ohjelmiston käyttö on sekä tehokasta että miellyttävää. Käytettävyyden merkitys on kasvanut, kun ohjelmistojen käyttäjiksi on tullut tietotekniikan ammattilaisten lisäksi kuluttajia. Kilpailukyvyn varmistamiseksi ohjelmistojen tulee olla käytettävän lisäksi myös saavutettavia ja niiden kokonaiskäyttökokemuksen on oltava hyvä. Saavutettavuus on myös lainmukainen vaatimus ohjelmistoille. Käytettävyyden testaaminen on perinteisesti käsityötä, jota suorittaa koulutettu käytettävyysasiantuntija tai käyttäjätutkija. Suuri osa huomiosta keskittyy käyttöliittymään, jonka kautta käyttäjä on vuorovaikutuksessa ohjelmiston kanssa. Käytettävyys on kontekstisidonnaista, joten osa testaamisesta tulee aina vaatimaan ihmisen analysointia. Osa testaamisesta voidaan kuitenkin automatisoida, ja näin vapauttaa tutkijan kapasiteettia sellaisiin osa-alueisiin, jotka vaativat kontekstin ja käyttäjän toiminnan ymmärtämistä ja tulkitsemista. Lisäksi tutkija pystyy keskittymään käytettävyysongelmia korjaavien muutosehdotusten kehittämiseen. Automaattisesti voidaan testata erityisesti käyttöliittymän yhdenmukaisuutta, käyttöliittymäelementtien havaittavuutta ja kerätä automaattisesti käyttäjän raportoimia ongelmakohtia. Tässä työssä pyritään vastaamaan kysymykseen, missä määrin käytettävyyden varmistamista pystytään automatisoimaan, ja esitellään tämän tiedon pohjalta rakennettu käytettävyyden automaattisen testaamisen työkalu, Usability Spy, jonka tarkoituksena on automatisoida käytettävyystesteihin osallistuvilta henkilöiltä koottavien käytettävyystestiaineiston keräystä. Työkalun toimintaympäristönä on DOS käyttöjärjestelmän päällä toimiva Microsoft Windows tai Microsoft Windows for Workgroups ikkunointiympäristön versio 3.1 tai uudempi.
  • Koivula, Kalle-Matti (2023)
    In this thesis we try to find the measurement accuracy of our dronebound wind measurement setup and if the quality of the measurements is high enough for operational usage. The thesis goes over the most important theoretical concepts concerning effects of wind in the boundary layer. In the thesis we analyze wind data gathered by a drone-bound anemometer, and introduce a direct method of measuring wind with a UAV. The data includes stationary wind data gathered at height of 30 metres, as well as vertical wind profiles to 500 metres above ground level. The data is compared to reference data from a 30 metre wind mast and automatic radiosoundings. The measurements were conducted in Jokioinen, Finland between the 2nd of September 2022 and 10th of October 2022. Total of 20 measurement flights were conducted, consisting of 14 stationary wind measurements and six wind profile measurements. We found out the stationary wind measurement quality to be comparable with earlier studies. The vertical wind profile measurements were found to be hard to analyze, as the reference measurement was not as compatible as we had hoped for. The difference between automatic radiosoundings and our profile measurements was distinctly greater than the difference between the stationary drone and wind mast measurements. Lastly some optimization and improvements to the measurement arrangement are discussed. The application of these improvements and modifications will be left as future endeavour for some willing individual.