Skip to main content
Login | Suomeksi | På svenska | In English

Browsing by Title

Sort by: Order: Results:

  • Seppälä, Minttu (2020)
    Observing stellar occultations is a great way to obtain information, such as astrometry and angular diameter, on objects that would be too faint to observe directly. Observing stellar occultations by minor bodies has become more common in recent years with the help of high-precision astrometry for around one billion stars from the Gaia Data Release 2. The Outer Solar System Origins Survey has provided high-precision positions for over eight-hundred transneptunian objects (TNOs). The goal of this survey was to study orbital distributions of TNOs to learn more about the early history of the Solar System. A secondary science goal was to allow occultation predictions to be made for the identified TNOs. A new statistical method was created to make occultation predictions. The goal was to make occultation predictions for the TNOs identified in the Outer Solar System Origins Survey. The method was tested on 2014 MU69 and it successfully found the three existing occultation events that occurred in 2017. The method was then applied to two TNOs observed in the Outer Solar System Origins Survey. The uncertainty limits for the shadow position on the Earth were larger than in the three existing predictions, but the overall shadow positions were very similar. The method was able to find a dozen of potential occultations for the two TNOs. However, the uncertainty limits for the shadow positions were very large, around the size of the Earth.
  • Keningi, Eino (2022)
    In little over a decade, cryptocurrencies have become a highly speculative asset class in global financial markets, with Bitcoin leading the way. Throughout its relatively brief history, the price of bitcoin has gone through multiple cycles of growth and decline. As a consequence, Bitcoin has become a widely discussed – and polarizing – topic on Twitter. This work studies whether the sentiment of popular Bitcoin-related tweets can be used to predict the future price movements of bitcoin. In total, seven different algorithms are evaluated: Vector Autoregression, Vector Autoregression Moving-Average, Random Forest, XGBoost, LightGBM, Long Short-Term Memory, and Gated Recurrent Unit. By applying lexicon-based sentiment analysis, and heuristic filtering of tweets, it was discovered that sentiment-based features of popular tweets improve the prediction accuracy over baseline features (open-high-low-close data) in five of the seven algorithms tested. The tree-based algorithms (Random Forest, XGBoost, LightGBM) generally had the lowest prediction errors, while the neural network algorithms (Light Short-Term Memory and Gated Recurrent Unit) had the poorest performance. The findings suggest that the sentiment of popular Bitcoin-related tweets can be an important feature in predicting the future price movements of bitcoin.
  • Duplouy, Lucie (2016)
    Found together in nature, the isolation of rare earth elements (REEs) has always been challenging. Their separation usually involves processes including solvent extraction, precipitation and ion exchange. Zeolites are well-known for their ion exchange capabilities that would be potentially applicable for the REEs separation. This study primarily investigated the ion exchange behavior of REEs onto several types of zeolites, namely the Ferrierite, the Faujasite and the Linde Type L. These zeolites were chosen because of their altered framework type and controllable charge density through synthesis. The commercially available zeolites were systematically characterized by means of X-Ray Diffraction (XRD), Energy Dispersive X-ray spectroscopy (EDX) and Attenuated Total Reflectance Fourier Transform Infrared spectroscopy (ATR-FTIR). Lanthanum was chosen as the model REE. The pH-uptake behavior of La on zeolites was studied extensively. Results showed that among all our selected zeolites, only LTL gave reasonable amount of La ion exchange capacity (0.16 mmol/g at pH 3). Subsequently, the ion exchange isotherm of La on LTL was described at pH 3. The isotherm followed Langmuir type with a maximum capacity of 0.25 mmol/g. An equimolar ternary mixture of lanthanum, neodymium and dysprosium was used to test the selectivity of LTL zeolite towards different REEs. Results suggested that the uptake sequence followed La > Nd > Dy, which indicated the decrease of capacity with increased atomic number. Almost all the REEs could be leached off from the REE loaded zeolites with a pH 1.51 nitric acid solution.
  • Albadry, Dalal (2014)
    Bierol, tetramer lignin model compound was synthesized from guaicylglycerol-ß-guaiacyl ether (Erol) by oxidative dimerization using Horseradish peroxidase (HRP) / H2O2 biomimetic oxidation system. Starting material Erol, a (β-O-4) type lignin model compound was synthesized in 80% overall yield in six steps from acetovanillone. The reactions progresses were associated with thin layer chromatography (TLC) and the purity of each product was obtained by High Performance Liquid Chromatography (HPLC) analysis techniques. The 1H and 13C NMR characteristics of the resulting products were studied using 1H and 13C NMR spectra as well as 2D NMR, in particular (Heteronuclear Single Quantum Coherence) HSQC. The objectives of this series of investigation was three-fold: (a) to synthesize dimeric lignin model compounds (ß-O-4), (b) to synthesize tetrameric lignin model compounds, (c) to study 1H and 13C NMR and 2D NMR spectroscopic characteristics of compounds. The research conducted in this project was successful as the model compounds were synthesized and characterized. Accordingly, the results were discussed in terms of signal assignments that confirmed the molecules structures. The general knowledge of lignin, e.g. chemical structure of lignin, the synthetic methods for synthesizing (ß-O-4) type model compounds, the different methods for lignin oxidative degradation and the various methods of lignin analysis were reviewed.
  • Featherstone, Graham Anthony (2016)
    In this thesis, the synthesis of titanium dioxide and sodium doped titanium dioxide nanofibers was undertaken through the use of the relatively new methods of solution blow spinning and electroblowing. These techniques are initially compared to other modern methods of nanofiber synthesis such as electrospinning, drawing, melt spinning and dry spinning. These techniques are evaluated based on production rate and the diameter of the formed nanofibers. This comparison shows that electroblowing and solution blow spinning are efficient high throughput methods for the formation of unordered nanofiber mats with diameters similar to those obtained in electrospinning. The formation of titanium dioxide nanofibers due to its role as a catalyst was of particular interest. Solution blow spinning and electroblowing are methods which employ the use of a high-velocity gas in order to stretch and elongate a viscous polymeric solution. While solution blow spinning relies entirely on the use of high-velocity gas, electroblowing additionally charges the elongated nanofibers which aids in the collection of the spun fibers, and this also affects the morphology of the final nanofibers. Inorganic nanofibers are obtained by adding a titanium dioxide precursor to the polymeric solution, which, upon calcination, forms solid titanium dioxide nanofibers. Dopants may also be added to the solution which allows for the formation of doped titanium dioxide nanofibers. Various solution and process parameters were studied in-depth in order to develop a full understanding of their effects on the diameters of the synthesized nanofibers. These parameters include the pressure of the gas, the feed-rate of the polymeric solution through the needle tip, the voltage applied to the needle tip, the concentration of the polymer and the distance from needle to the collector. After process and solution optimization, production rates of 0.39 g/h and 0.55 g/h were obtained for the titanium dioxide and sodium doped titanium dioxide nanofibers, respectively. With these optimized parameters, the average titanium dioxide fiber diameters measures 182 nm while the average sodium doped titanium dioxide diameters measured 184 nm. Crystallization studies were also conducted on the calcinated nanofibers. Both high-temperature in situ XRD studies as well as room temperature measurements on calcinated samples were done in order to cross-compare results and eliminate any errors associated with each individual method. The titanium dioxide nanofibers demonstrated a very defined crystallinity in which the fibers shifted from anatase to a predominantly rutile phase between the temperatures of 410 to 1050 °C. However, the sodium doped fibers demonstrated a mixed phase crystallinity in which no crystal structure was discernible.
  • Leivo, Marcus Johannes (2020)
    Problems which ask to compute an optimal solution to its instances are called optimization problems. The maximum satisfiability (MaxSAT) problem is a well-studied combinatorial optimization problem with many applications in domains such as cancer therapy design, electronic markets, hardware debugging and routing. Many problems, including the aforementioned ones, can be encoded in MaxSAT. Thus MaxSAT serves as a general optimization paradigm and therefore advances in MaxSAT algorithms translate to advances in solving other problems. In this thesis, we analyze the effects of MaxSAT preprocessing, the process of reformulating the input instance prior to solving, on the perceived costs of solutions during search. We show that after preprocessing most MaxSAT solvers may misinterpret the costs of non-optimal solutions. Many MaxSAT algorithms use the found non-optimal solutions in guiding the search for solutions and so the misinterpretation of costs may misguide the search. Towards remedying this issue, we introduce and study the concept of locally minimal solutions. We show that for some of the central preprocessing techniques for MaxSAT, the perceived cost of a locally minimal solution to a preprocessed instance equals the cost of the corresponding reconstructed solution to the original instance. We develop a stochastic local search algorithm for MaxSAT, called LMS-SLS, that is prepended with a preprocessor and that searches over locally minimal solutions. We implement LMS-SLS and analyze the performance of its different components, particularly the effects of preprocessing and computing locally minimal solutions, and also compare LMS-SLS with the state-of-the-art SLS solver SATLike for MaxSAT.p
  • Partovi, Fariba (2021)
    Utilization of pesticides in the modern agriculture is often indispensable for gaining good crops. However, pesticides are abundantly being used in too hight quantities which leads to potential health risks for the consumers. Currently there are no pre-screening methods for monitoring the levels of pesticides in food, but only a negligible small percentage of all goods are being tested using the laborious standardized methods. This master’s thesis is an investigation, that was carried out in the wet laboratory of KARSA Oy Ltd, on 10 different pesticides: Glyphosate, Thiabendazole, 2-phenylphenol, Chlorpyrifos, Fludioxonil, Chlormequat, Bupirimate, Diflubenzuron, Fenpyrazamine and 2,4-dichlorophenoxyacetic acid. Pesticides were ionized using straight radiation chemical ionization (SRCI) in positive and negative modes without any added reagent and also using bromide, nitrate, acetonylacetone and acetone as reagents. Charged target molecules and adducts were detected using Thermo fisher Iontrap/Orbitrap (LTQ Orbitrap velos pro upgraded) mass spectrometer. After the initial method development and scoping measurements pesticides were studied both individually and as a mixture of all 10 pesticides. Sample solutions were first injected with syringe so that the solvent and targets evaporated at the same time inside the desorber heating block of SRCI inlet. In these syringe injection measurements, the desorber temperature was set at 150 °C. Mass range at 125–750 has been used for all the pesticides except for Chlormequat (100–750). After the syringe injection measurements, the mixture of 10 pesticides was analysed by TCM filters. Target solutions of 1 µl volume were placed on filters and after the solvents had evaporated the filters were heated from room temperature to 245°C using the same setup as with the syringe injections. In conclusion, with syringe injections 7 pesticides out of 10 were detected using positive and negative mode without any added reagent. The highest target intensities were recorded from TCM filters. Overall, applying the SRCI-Orbitrap setup for pesticide pre-screening from target solutions resulted in the detection of 9 pesticides out of 10.
  • Timofeeva, Natalia (2019)
    This thesis attempts to show the advantages of Monte Carlo method in pricing and hedging exotic options. The popularity of exotic options increased recently, mostly due to their almost unlimited flexibility and adaptability to any circumstances. The negative side of exotic options is their complexity. Due to that many exotic options does not have analytic solutions. As a result numerical solutions are a necessity. The Monte Carlo method of simulations is very common method in computational finance. Monte Carlo method is based on the analogy between probability and volume. Starting point in pricing and hedging options with Monte Carlo method is stochastic differential equation based on Brownian motion in the Black-Scholes world. The fair option value in the Black-Scholes world is the present value of the expected payoff at expiry under a risk-neutral assumptions. The analysis start from the case of the simple European options and continue with introducing different kinds of exotic options. The dynamic hedging idea is used to derive the Black-Scholes Partial Differential Equation. The numerical approximation of the stochastic differential equation is derived through the lognormal asset price model. The Monte Carlo algorithms are constructed for pricing and delta hedging and then implemented to MATLAB. For generating Monte Carlo simulations is used N(0,1) pseudo-random generator. The analysis is limited to the cases of simple Barrier options, which are one of the most known and used type of the exotic options. Barrier options are path dependent options, which implies that the payoff depends on the path followed by the price of the underlying asset, meaning that barrier options prices are especially sensitive to volatility. That is why, we also introduce the variance reduction techniques by antithetic variates. For hedging barrier options were chosen the dynamic delta-hedging and static hedging strategies. All calculations and figures in the examples were made in MATLAB.
  • Pirilä, Pauliina (2024)
    This thesis discusses short-term parking pricing in the context of Finnish shopping centre parking halls. The focus is on one shopping centre located in Helsinki where parking fees are high and there is a constant need for raising the prices. Therefore, it is important to have a strategy that maximises parking hall income without compromising the customers' interest. If the prices are too high, customers will choose to park elsewhere or reduce their parking in private parking halls. There is a lot of competition with off-street parking competing against on-street parking and access parking, not to mention other parking halls. The main goal of this thesis is to raise problems with parking pricing and discuss how to find the most beneficial pricing method. To achieve this, this thesis project conducted an analysis on one Finnish shopping centre parking hall data. This data was analysed to discover the average behaviour of the parkers and how the raised parking fees affect both the parker numbers and the income of the parking hall. In addition, several pricing strategies from literature and real-life examples were discussed and evaluated, and later combined with the analysis results. The results showed that there are some similarities with results from literature but there were some surprising outcomes too. It seems that higher average hourly prices are correlated with longer stays, but still the parkers who tend to park longer have more inelastic parking habits than those who park for shorter durations. The calculated price elasticity of demand values show that compared to other parking halls, parking is on average more elastic in the analysed parking hall. This further emphasises the importance of milder price raises at least for the shorter parking durations. Moreover, there are noticeable but explainable characteristics in parker behaviour. Most of the parkers prefer to park for under one hour to take advantage of the first parking hour being free. This leads to profit losses in both the shopping centre and parking hall income. Therefore, a dynamic pricing strategy is suggested as one pricing option, since it adjusts the prices automatically based on occupancy rates. Although there are some challenges with this particular method, in the long run it could turn out to be the most beneficial for both the parking hall owners and the parkers. To conclude, choosing a suitable pricing strategy and model for a parking hall is crucial and the decisions should be based on findings from data.
  • Kuronen, Toini (2016)
    This study focuses on looking at different factors that have been shown to explain why certain farms are vulnerable to crop-raiding by primates. The type of wildlife value orientations present among the study sample are also looked into in this study. Furthermore, how these wildlife value orientations affect how the community members perceive and react to human-wildlife conflicts is also discussed. The data collection was conducted in May and June 2015 around Ngangao indigenous cloud forest in the Taita Hills, southeast Kenya. The study is both qualitative and quantitative. Semi-structured household-based questionnaires were used in the primary data collection. Also, 11 local experts were interviewed and two workshops were held where the participants were given topics about wildlife issues and solutions to discuss about. Statistical analysis as well as spatial analysis using GIS were performed in this study. The findings of the study are that the closer a farm is to the forest boundary and the less neighbouring farms there are between the farm and the forest, the more vulnerable that farm is to crop-raiding by Sykes' monkeys. The study could not prove that a specific type of food crop grown in a farm or the type of land use between the farmland and the forest boundary is explaining vulnerability to crop-raiding by primates. Moreover, strong determinants that explain the vulnerability of a certain farm to crop-raiding by vervet monkeys, yellow baboons or bush-babies were not found in this study. The majority of the studied households practice subsistence farming as their main livelihood. Therefore, crop-raiding by wildlife, such as primates, is a severe threat to the food security and livelihoods of local households. The study points out that a majority of the study participants perceive wildlife in a materialistic way, either as threats or as benefits. A smaller share of respondents represent a harmonius wildlife value orientation. Because majority of the local community is likely to represent similar wildlife value orientations, crop-raiding by primates is perceived as a significant problem. Some community members are believed to have reacted to human-primate conflicts by deliberately setting forest fires in the forests of the Taita Hills to deter the marauding primates away. This study suggests that community representatives, local wildlife management and officials should collectively discuss and address the issue of human-primate conflict in the indigenous forests of the Taita Hills so that the community level perceptions are not disregarded in wildlife management. Possible solutions to human-primate conflicts are, for example, providing the most vulnerable households with compensation, incentives or food relief and tools to practice alternative livelihoods. Moreover, relocation of certain groups of primates and planting wild fruit trees inside indigenous forests could ease the problem. Community wildlife association could work as a platform to address the issue. Additionally, sensitizing the local communities about local nature and addressing the awareness gaps regarding problem wildlife reporting could increase the tolerance to wildlife damages at community level and also support the coexistence of humans and wildlife.
  • Tervo, Reetta (2020)
    Tutkielman tarkoituksena on perehdyttää lukija primitiivisten juurten ja neliönjäännösten teoriaan, minkä lisäksi esitellään niiden joitakin perusominaisuuksia. Primitiiviset juuret ja neliönjäännökset ovat osa matemaattisenlukuteorian alaa. Tutkielman alkuun esitellään Eulerin phi-funktio ja määritellään kokonaisluvun kertaluku, joita tarvitaan primitiivisten juurten määrittämisessä. Vuonna 1773 Euler antoi ensimmäisen todistuksensa sille, että jokaiselle alkuluvulle on olemassa primitiivisiä juuria, ja myöhemmin Legendre korjaili tässä todistuksessa ilmenneitä epäkohtia. Vuonna 1801 Gauss puolestaan otti suuren edistysaskeleen primitiivisten juurten perusteoriassa, kun hän onnistui osoittamaan kaikki ne luvut, joille on olemassa primitiivisiä juuria ja todistamaan, että nämä luvut ovat ainoat, joilla niitä on olemassa. Primitiivisten juurten sovelluskohteista perehdytään tässä tutkielmassa tarkemmin desimaaliekspansioon. Tutkielman neljännessä luvussa esitellään desimaaliekspansion määritelmä, minkä jälkeen tutkitaan rationaalilukujen päättymistä. Tämän jälkeen selvitetään primitiivisiä juuria apuna käyttäen, kuinka pitkä on alkuluvun käänteisluvun päättymättömän desimaaliekspansion toistuva periodi. Tutkielman viides luku pitää sisällään neliönjäännöksien perusominaisuuksia. Luvussa annetaan Eulerin kriteerille kaksi erilaista esitysmuotoa: ensimmäinen Legendren symbolilla ja toinen ilman. Tämä lause on neliönjäännösten resiprookkilain kannalta erittäin merkittävä ja tässä tutkielmassa sille esitetään myös kaksi todistusta, joista toisessa käytetään apuna primitiivisiä juuria. Vielä ennen neliönjäännösten resiprookkilakia esitellään Gaussin lemma, jota sovelletaan resiprookkilain todistuksessa. Viimeisessä luvussa pohditaan, kuinka primitiivisten juurten ja neliönjäännösten opetus soveltuu lukion matematiikan kursseille. Lukiossa on tällä hetkellä olemassa pitkän matematiikan kurssi, jossa tutustutaan lukuteoriaan. Tämä kurssi tulee uudistumaan merkittävästi uuden lukion opetussuunnitelman myötä vuonna 2021, minkä seurauksena tämän tutkielman aihealueet ovat paremmin sisällytettävissä siihen. Tulevia tutkimuksia varten olisi mielenkiintoista selvittää, kuinka vanhan opetussuunnitelman kurssin käsiteltävät aiheet ja vaikeustaso tulevat muuttumaan algoritmien ja ohjelmoinnin myötä.
  • Halkoaho, Johannes (2022)
    The primordial perturbations created by inflation in the early Universe are known to be able to produce significant amount of primordial black holes and gravitational waves with large amplitudes in some inflationary models. Primordial black holes are produced by primordial scalar perturbations and gravitational waves are partly primordial tensor perturbations and partly produced by scalar perturbations. In this thesis we review some of the current literature on the subject and discuss a few inflationary models that are capable of producing primordial scalar perturbations large enough to create a significant amount of primordial black holes. The main focus is on ultra-slow roll inflation with a concrete example potential illustrating the dynamics of the scenario followed by a briefer treatment of some of the alternative models. We start by explaining the necessary background theory for the understanding of the subject at hand. Then we move on to the inflationary models covered in this thesis. After that we explain the production of the primordial black holes and gravitational waves from scalar perturbations. Then we consider primordial black holes as a dark matter candidate and go through the most significant known restrictions on the existence of primordial black holes with different masses. We discuss some of the possible future constraints for the remaining possible mass window for which primordial black holes could explain all of dark matter. We then briefly discuss two planned space-based gravitational wave detectors that may be able to detect gravitational waves created by inflation.
  • Palmgren, Elina (2015)
    The phase of accelerating expansion of the early universe is called cosmological inflation. It is believed that the acceleration was driven by a scalar field called the inflaton, which in the simplest inflationary models was slowly rolling down its potential (slow-roll inflation). As an extension of these simple models, in this work we study a model in which the rolling of the inflaton field was fast (fast-roll inflation). From the inflationary phase we can detect density fluctuations which are usually thought to be created from quantum fluctuations of the inflaton field. In the case of the fast-rolling inflaton field, the evolution of the perturbations may differ significantly from the simpler slow-roll models. In addition to the inflaton, there might have also been other quantum fields in the early universe acting on the perturbation evolution. In this work, we study a case in which the perturbations are partly created from the inflaton field fluctuations and partly from another scalar field named the curvaton. This kind of scenario is called the mixed model. In the investigation of the early universe, it is essential to be able to estimate to what extent the observed perturbations tell us directly of the inflaton dynamics – and not of other phenomena. From the inflationary model-building point of view, freeing the inflaton from often somewhat unnatural observational constraints gives us new possibilities to develop more plausible models of inflation. In this thesis, we introduce briefly the simplest inflation and curvaton models and study analytically the perturbation evolution in the mixed model and fast-roll inflation. The research part of the thesis contains a numerical investigation of the perturbation evolution, as well as the inflaton and curvaton dynamics, in these models.
  • Kainberg, Alexander (2012)
    I slutet av 1700-talet gissade Gauss och Legendre att lim_{x→∞} \frac{π (x) log (x)}{x} =1, där π (x) är antalet primtal som är mindre än eller lika med x. Hadamard och de la Vall'ee Poussin bevisade påstående oberoende av varandra år 1896 och resultatet kallas numera primtalssatsen. Efter detta har satsen bevisats på både elementära sätt (Selberg & Erdõs, 1949) och med hjälp av komplexanalys (Newman, 1980). I denna avhandling kommer vi att presentera ett analytiskt bevis av PTS. I beviset kommer vi att utnyttja Riemanns zetafunktion och dess egenskaper. I kapitel 3 behandlar vi komplexanalys. Vi diskuterar bl.a. Eulers gammafunktion, Riemanns zetafunktion och Dirichletserier. Det är en naturlig fortsättning till PTS att diskutera zetafunktionens nollställen, och därför bevisar vi Hardys sats. Efter detta kommer vi att presentera den ökända Riemannhypotesen. Gauss gissade att den logaritmiska integralen \li (x) :=∈t_2^x \frac{dt}{log t} skulle approximera π (x) mycket bra. Vi definierar den s.k. restfunktionen som r(x):=π (x) - \li (x). Vi avslutar avhandlingen genom att bevisa att om det för varje \epsilon >0 gäller att |r(x)|<x^{\frac{1}{2} +\epsilon} så är Riemannhypotesen sann.
  • Chelak, Ilia (2024)
    Recently, 3D reconstruction has become a popular topic due to applications in Virtual Reality, Augmented Reality, and historical heritage preservation. Yet, high-quality reconstruction is not available to the general public because of the cost of laser scanners. The goal of this thesis is to make the democratization of 3D reconstruction closer with the use of photogrammetry (reconstruction from multi-view images.) However, current approaches are very slow and tend to oversmooth the geometry. Our method involves learning the scene via a neural representation by taking posed multi-view images as input. We note that state-of-the-art (SOTA) approaches rely on traditional Structure-from-Motion (SfM) algorithms to extract camera poses. We also observe that SfM can generate a coarsely correct mesh for the underlying object. Nevertheless, SOTA techniques start training the neural representation from a sphere. Therefore, we propose a novel initialization method that takes the mesh obtained from SfM and initializes the neural representation from it. We validate our approach through extensive experiments on a widely used multi-view stereo DTU dataset. We show that our method outperforms both traditional and SOTA neural techniques in terms of reconstruction quality. It manages to learn the underlying geometry and recover small details like cracks and dents. We also show that it speeds up convergence by 4 times. All the datasets, reconstructed meshes, and learned model weights are available at this link.
  • Jälkö, Joonas (2017)
    This thesis focuses on privacy-preserving statistical inference. We use a probabilistic point of view of privacy called differential privacy. Differential privacy ensures that replacing one individual from the dataset with another individual does not affect the results drastically. There are different versions of the differential privacy. This thesis considers the ε-differential privacy also known as the pure differential privacy, and also a relaxation known as the (ε, δ)-differential privacy. We state several important definitions and theorems of DP. The proofs for most of the theorems are given in this thesis. Our goal is to build a general framework for privacy preserving posterior inference. To achieve this we use an approximative approach for posterior inference called variational Bayesian (VB) methods. We build the basic concepts of variational inference with certain detail and show examples on how to apply variational inference. After giving the prerequisites on both DP and VB we state our main result, the differentially private variational inference (DPVI) method. We use a recently proposed doubly stochastic variational inference (DSVI) combined with Gaussian mechanism to build a privacy-preserving method for posterior inference. We give the algorithm definition and explain its parameters. The DPVI method is compared against the state-of-the-art method for DP posterior inference called the differentially private stochastic gradient Langevin dynamics (DP-SGLD). We compare the performance on two different models, the logistic regression model and the Gaussian mixture model. The DPVI method outperforms DP-SGLD in both tasks.
  • Bärlund-Vihtola, Nina (2020)
    Henkilötietojen suoja eli tietosuoja on Euroopan unionin perusoikeuskirjassa vahvistettu perusoikeus. Yksityishenkilöiden tietosuoja vahvistui entisestään, kun vuonna 2018 alettiin kaikissa EU:n jäsenmaissa soveltaa Euroopan parlamentin ja neuvoston asetusta (EU) 2016/679 luonnollisten henkilöiden suojelusta henkilötietojen käsittelyssä sekä näiden tietojen vapaasta liikkuvuudesta ja direktiivin 95/46/EY kumoamisesta, lyhyemmin EU:n yleistä tietosuoja-asetusta. Asetusta täydennettiin myöhemmin kansallisella tietosuojalailla 1050/2018. Organisaatio, joka omistaa henkilörekisterin, on rekisterinpitäjä, ja henkilö, jonka tiedot ovat henkilörekisterissä, rekisteröity. Tietosuoja-asetuksessa on määritetty rekisterinpitäjän velvollisuudet, joiden avulla tämän on vastattava siitä, että rekisteröidyn oikeudet toteutuvat. Lisäksi rekisterinpitäjällä on osoitusvelvollisuus siitä, että henkilötietojen käsittely tapahtuu tietosuoja-asetuksen käsittelyperiaatteiden mukaisesti, sekä ilmoitusvelvollisuus kansalliselle valvontaviranomaiselle tilanteissa, joissa on tapahtunut henkilötietojen tietoturvaloukkaus. Tietosuoja-asetuksella on vaikutuksia sovelluskehitykseen. Asetus velvoittaa huolehtimaan sisäänrakennetusta tietosuojasta (Privacy by Design) eli tietosuojaa tuottavien toiminnallisuuksien toteutuksesta tietojärjestelmiin alusta alkaen. Asetus on pitkä ja hankala, joten ollut organisaatioille työlästä havaita siitä kaikki velvoitteet. Tässä tutkielmassa on pyritty vastaamaan tuohon ongelmaan etsimällä asetuksesta sisäänrakennetun tietosuojan vaatimukset ja kiinnittämällä ne TOGAF-kokonaisarkkitehtuurikehykseen. Tämän pohjalta on toteutettu sovelluskehityksen tietosuojaohjeistuksen runko, jota kehittämällä on organisaatiolle luotavissa toimiva ohjeistus sisäänrakennetun tietosuojan rakentamisen tueksi.
  • Ghasemi, Mandana (2019)
    Over the last years, Location-Based Services (LBSs) have become popular due to the global use of smartphones and improvement in Global Positioning System (GPS) and other positioning methods. Location-based services employ users' location to offer relevant information to users or provide them with useful recommendations. Meanwhile, with the development of social applications, location-based social networking services (LBSNS) have attracted millions of users because the geographic position of users can be used to enhance the services provided by those social applications. Proximity detection, as one type of location-based function, makes LBSNS more flexible and notifies mobile users when they are in proximity. Despite all the desirable features that such applications provide, disclosing the exact location of individuals to a centralized server and/or their social friends might put users at risk of falling their information in wrong hands, since locations may disclose sensitive information about people including political and religious affiliations, lifestyle, health status, etc. Consequently, users might be unwilling to participate in such applications. To this end, private proximity detection schemes enable two parties to check whether they are in close proximity while keeping their exact locations secret. In particular, running a private proximity detection protocol between two parties only results in a boolean value to the querier. Besides, it guarantees that no other information can be leaked to the participants regarding the other party's location. However, most proposed private proximity detection protocols enable users to choose only a simple geometric range on the map, such as a circle or a rectangle, in order to test for proximity. In this thesis, we take inspiration from the field of Computational Geometry and develop two privacy-preserving proximity detection protocols that allow a mobile user to specify an arbitrary complex polygon on the map and check whether his/her friends are located therein. We also analyzed the efficiency of our solutions in terms of computational and communication costs. Our evaluation shows that compared to the similar earlier work, the proposed solution increases the computational efficiency by up to 50%, and reduces the communication overhead by up to 90%. Therefore, we have achieved a significant reduction of computational and communication complexity.
  • Leinonen, Juho (2017)
    The computer science education research field studies how students learn computer science related concepts such as programming and algorithms. One of the major goals of the field is to help students learn CS concepts that are often difficult to grasp because students rarely encounter them in primary or secondary education. In order to help struggling students, information on the learning process of students has to be collected. In many introductory programming courses process data is automatically collected in the form of source code snapshots. Source code snapshots usually include at least the source code of the student's program and a timestamp. Studies ranging from identifying at-risk students to inferring programming experience and topic knowledge have been conducted using source code snapshots. However, replicating source code snapshot -based studies is currently hard as data is rarely shared due to privacy concerns. Source code snapshot data often includes many attributes that can be used for identification, for example the name of the student or the student number. There can even be hidden identifiers in the data that can be used for identification even if obvious identifiers are removed. For example, keystroke data from source code snapshots can be used for identification based on the distinct typing profiles of students. Hence, simply removing explicit identifiers such as names and student numbers is not enough to protect the privacy of the users who have supplied the data. At the same time, removing all keystroke data would decrease the value of the data significantly and possibly preclude replication studies. In this work, we investigate how keystroke data from a programming context could be modified to prevent keystroke latency -based identification whilst still retaining valuable information in the data. This study is the first step in enabling the sharing of anonymized source code snapshots. We investigate the degree of anonymization required to make identification of students based on their typing patterns unreliable. Then, we study whether the modified keystroke data can still be used to infer the programming experience of the students as a case study of whether the anonymized typing patterns have retained at least some informative value. We show that it is possible to modify data so that keystroke latency -based identification is no longer accurate, but the programming experience of the students can still be inferred, i.e. the data still has value to researchers.
  • Rantaniemi, Eero (2024)
    Tässä tutkielmassa tutustutaan proäärellisiin ryhmiin, siis topologisiin ryhmiin, jotka ovat isomorfisia äärellisten topologisten ryhmien muodostaman inverssisysteemin rajan kanssa. Tutkielman alussa esitetään topologisten ryhmien yleistä teoriaa, sekä tutustutaan inverssisysteemeihin yleisesti esittämällä kokoelma näiden ominaisuuksia. Tämän jälkeen tutkielmassa siirrytään käsittelemään proäärellisiä ryhmiä ja ja esitetään tärkeä karakterisaatio, jonka mukaan topologinen ryhmä on proäärellinen jos ja vain jos se on kompakti ja täysin epäyhtenäinen. Tästä seuraa Blairen lauseen nojalla, että jokainen proäärellinen ryhmä on Blairen avaruus. Tutkielman lopuksi käsitellään proäärellisiä täydellistymiä, erityisesti kokonaislukujen pro-p täydellistymämme, siis p-adisille luvuille, annetaan oma kappaleensa, jossa esitetään näiden konstruktio inverssisysteemin rajana sekä äärettömän pitkinä luonnollisina lukuina, sekä esitetään näiden ominaisuuksia. Näistä tärkeimpänä esitetään Henselin lemma, jonka avulla p-adisten lukujen polynomille löydetään juuri, kunhan sille on annettu riittähän hyvä arvio. Tällä tuloksella on käyttöä myös modulaariaritmeriikassa. Viimeisenä tutkielmassa esitetään kokonaislukujen proäärellinen täydellistymä