Skip to main content
Login | Suomeksi | På svenska | In English

Browsing by Title

Sort by: Order: Results:

  • Pylvänäinen, Annika (2012)
    Tämän Pro gradun aiheena on satunnaismatriisien ominaisarvojen jakautuminen ja jakauman soveltaminen. Keskitytään erityisesti gaussisiin matriisiensembleihin, toisin sanoen matriisikokoelmiin, joiden alkiot noudattavat normaalijakaumaa. Tämän jakaumaoletuksen pätiessä teoriaa sovelletaan dffuusio-MRI tutkimukseen. Ensimmäisessä luvussa tarkastellaan matriisien ominaisuuksia, jotka ovat keskeisessä roolissa satunnaismatriisien teoriassa. Määritellään neliömatriisin ominaisuuksia kuten matriisin neliömuoto ja determinantti. Määritellään lisäksi matriisiarvoinen satunnaismuuttuja ja sen seurauksena keskitytään satunnaismatriiseihin. Todennäköisyys on keskeinen työkalu satunnaisuutta käsiteltäessä ja määritelläänkin todennäköisyysteorian peruselementtejä. Niiden avulla voidaan laskea satunnaismatriisin multinormaalijakauma sekä sen ominaisarvot ja -vektorit. Luvussa 2 määritellään Wignerin reaalinen symmetrinen- ja Wignerin hermiittinen matriisi. Perehdytään ennen kaikkea gaussiseen ortogonaaliseen (GOE)- ja gaussiseen unitaariseen matriisiensembleen (GUE), jotka ovat Wignerin matriisien erikoistapauksia. Tarkastellaan gaussisten matriisien jakaumaa ja erityisesti lasketaan matriisin ominaisarvojen jakauma. Se on tämän Pro Gradun keskeisempiä tuloksia ja sitä voidaan luvussa 3 soveltaa myös magneettikuvauksen teoriaan. Määritetään lisäksi Mehtan ja Selbergin integraalit, joiden avulla voidaan määrittää jakauman normalisointivakio. Lopuksi tarkastellaan diffuusiotensori- ja diffuusiopainotteista magneettikuvausta. Kuvataan ensin veden di_uusiota toisen asteen tensoreiden ja diffuusiofunktion avulla. Tämä on kolmiulotteinen malli, joka kuvaa diffuusion suuntaa kudoksessa. Monimutkaisempien diffuusioprofiilien, kuten kudosten hienorakenteiden sekä kuitujen leikkauskohtien tarkastelemiseen tarvitaan korkeamman asteen tensoreita. Tutustutaan niiden käyttöön sekä käytön vaatimiin rajoituksiin. Tarkastellaan sekä vektorin että tensorin jakaumia. Määritellään lisäksi rajoitteet, jotka vaaditaan algebrallisten ja geometristen ominaisuuksien säilymiseen muuntautuessa vektori- ja tenroriarvoisten muuttujien välillä. Lasketaan myös jakauman normalisointivakio. Lopuksi tarkastellaan isotrooppisen tensorin ominaisarvojen jakaumaa.
  • Telivuo, Suvi (2018)
    Taloudellisten, institutionaalisten ja teknologisten ympäristöjen kiihtyvä muutostahti on luonut tarpeen tehdä oikeita valintoja menestymisen ja kehityksen takeeksi. Parhaimman valinnan tekee henkilö, jolla on eniten tietoa ja varmuutta tiedon paikkansapitävyydestä. Vaihtoehtoisesti päätöksenteon epävarmuustekijöitä voidaan hallita eliminointimenetelmillä, joiden hyödyntäminen voi myös johtaa parempiin päätöksiin. Epävarmuuden minimoimisen edellytyksenä on niin ikään pohjatietojen parantaminen. Tästä tarpeesta ovat nousseet tiedonlouhintamenetelmät. Tiedonlouhintamenetelmiä on kehittynyt valtava määrä vastaamaan kysynnän luomia tarpeita. Päätöksentekopuu on eräs tällainen analysointimenetelmä ja päätöksentekopuun pohjalta on luotu koneoppimismenetelmä satunnaismetsä. Satunnaismetsä on tutkimusten mukaan tällä hetkellä paras saatavilla oleva luokittelumenetelmä ja valikoitunut tämän tutkielman aiheeksi. Luvussa 2 luomme pohjaa satunnaismetsä-menetelmän ymmärtämiseksi. Lähdemme liikkeelle koneoppimisesta ja datalouhintamenetelmistä, joilla alustamme päätöksentekopuutyökalun. Käy ilmi, että on olemassa luokittelu-, regressio- ja luokitteluregressiopuita, ja että tässä tutkielmassa keskitymme luokittelupuihin. Tämän jälkeen esittelemme päätöksentekopuun metodologiaa. Luvussa 3 esittelemme tutkielman kannalta päätöksentekopuiden tärkeimmät validointimenetelmät, sillä datalouhinnassa analysointimenetelmien validoiminen on yhtä tärkeää kuin itse analysoiminen. Esittelemme mallien validointiin liittyviä käsitteitä, kuten tarkkuus, yleistysvirhe ja ylisovittuminen. Käymme läpi yleisimpiä tapoja validoida malleja, sekä näytämme esimerkkien kautta työkalut, joita käytämme tutkielmassa. Näitä ovat tarkkuus, väärinluokittelumatriisi, ROC-, kumulatiivinen saanti- ja nostokäyrä. Luvussa 4 esittelemme satunnaismetsä-koneoppimismenetelmän. Käymme ensiksi läpi joukko-oppimisen metodologiaa, jonka jälkeen käsittelemme satunnaismetsän algoritmin. Osoitamme teoreettisesti, miksi satunnaismetsä on parempi luokittelija kuin esimerkiksi päätöksentekopuu näyttämällä, että satunnaismetsän puumäärän kasvaessa satunnaismetsän yleistysvirhe suppenee kohti nollaa. Analysoimme teoreettisesti satunnaismetsän hyötyjä ja haittoja. Satunnaismetsän hyötyjä ovat sen tarkkuus, nopeus, ymmärrettävyys, toimivuus valtavilla datamäärillä, sekä kykeneväisyys analysoida tietojoukon merkittävimpiä muuttujia. Haittoja ovat, että satunnaismetsä ei suoriudu yhtä hyvin regressio-ongelmissa kuin esimerkiksi logistinen regressiomalli, sekä huono sovellettavuus pieniin tietojoukkoihin. Luvussa 5 sovellamme opittuja taitoja suppeaan tietojoukkoon. Tarkoituksenamme on arvioida RStudion ja SAS Enterprise Minerin satunnaismetsä-pakettien toimivuutta tunnetulla syötejoukolla. Analysoimme satunnaismetsäin suoriutumista luokittelutehtävässä ja vertailemme tuloksia päätöksentekopuuhun ja regressiomalliin. Hyödynnämme luvussa 3 opittuja validointimenetelmiä. Käy ilmi, että RStudion ja SAS Enterprise Minerin satunnaismetsä-paketit toimivat hyvin, ja että satunnaismetsä suoriutuu pienenkin tietojoukon luokittelussa malleista parhaiten. Luvussa 6 sovellamme satunnaismetsää yrityksen tarjoamaan haasteeseen, jossa tarkoitus on selittää ja ennustaa uusien asiakkaiden tietyn asiakassegmentin asiakasvaihtuvuutta. Käytämme yhtiön tarjoamia tietokantoja ja SAS Enterprise Miner-työkalua. Suoritamme vertailun satunnaismetsä- ja päätöksentekopuu-mallien välillä käyttämällä luvussa 3 esiteltyjä validointimenetelmiä ja analysoimme tulokset. Käy ilmi, että ROC-käyrien ja tarkkuuden perusteella satunnaismetsä suoriutuu sekä luokittelussa että ennustamisessa paremmin kuin päätöksentekopuu. Luvussa 7 pohdimme, millaisissa puitteissa satunnaismetsä soveltuu yrityksen liiketoimintaprosessiin. Käymme läpi vaatimuksia, joita satunnaismetsän soveltaminen asettaa, sekä mitä lisäarvoa satunnaismetsä menetelmänä tuo yritykselle. Tulos on, että satunnaismetsä soveltuu hyvin yrityksille, jotka hyödyntävät SAS-työkaluja ja tuo lisäarvoa analysointitehtäviin olemalla ymmärrettävä malli, mutta kuitenkin monipuolinen, nopea ja tarkka.
  • Ahola, Johanna (2017)
    Tässä työssä tutkittiin poly(2-(dimetyyliamino)etyyli metakrylaatin) (PDMAEMA) oksastamista eri savipartikkelien pintaan kontrolloidulla radikaalipolymerointimenetelmällä ja sen vaikutusta partikkelien kykyyn absorboida tuoksumolekyylejä (tässä appelsiiniöljy) verrattuna puhtaisiin saviin. Lisäksi pyrittiin selvittämään, miten polymerointi vaikuttaa tuoksun pidättymisaikaan savimateriaaleissa ja onko tuoksun vapautuminen polymeroinnin jälkeen kontrolloitua. Ennen tuoksumolekyylien imeytymis- ja vapautumistutkimusta tavoitteena oli löytää toistettava pintainitioitu atominsiirtoradikaalipolymerointi-synteesimenetelmä (SI-ATRP) savi/polymeerikomposiittien valmistamiseksi. Savi/PDMAEMA-komposiitit valmistettiin syntetisoimalla ensin savi/aminosilaanikomposiitteja, joihin liitettiin initiaattori 2-bromoisobutyryylibromidi. DMAEMA:n polymerointi suoritettiin savi/initiaattori-komposiittien pintaan ‘grafting from’ -tekniikalla käyttäen ATRP-menetelmää. Savina käytettiin montmorilloniittia, halloisiittia ja wollastoniittia, joista montmorilloniitti- ja halloisiitti/PDMAEMA-komposiittien syntetisoinnissa onnistuttiin. Lähtöainesavien rakenne ja dimensiot tutkittiin kuvaamalla ne kenttäemissiopyyhkäisy-elektronimikroskoopilla. Väli- ja lopputuotteiden rakenteet karakterisoitiin IR- ja 1H-NMR-spektrometrisesti sekä termogravimetrisesti. Imeytyneen/vapautuneen appelsiiniöljyn määrä ja öljyn pysyminen savissa ja savi/PDMAEMA-komposiiteissa todennettiin termogravimetrisesti (dynaamisilla ja isotermisilla TGA-määrityksillä). Polymeerien ketjunpituudet ja polydispersiteetit määritettiin kokoekskluusiokromatografisesti (GPC). Savi/PDMAEMA-komposiittien valmistamiseksi löydettiin toimiva SI-ATRP-menetelmä, ja polymeeriketjujen kiinnittyminen saven pintaan todistettiin. Tutkimus osoitti PDMAEMA-oksastuksen vaikuttavan appelsiiniöljyn imeytymis- ja vapautumisominaisuuksiin siten, että öljyä imeytyi oksastettuun komposiittiin enemmän ja se pysyi komposiittimateriaalissa kauemmin verrattuna puhtaisiin saviin. Vaikka tutkimuksen tulokset osoittivat polymeroinnin merkittävän hyödyn tuoksuominaisuuksien parantamisessa, täysin kontrolloitua vapautumissysteemiä ei onnistuttu luomaan.
  • Lee, Hyeongju (2021)
    The number of IoT and sensor devices is expected to reach 25 billion by 2030. Many IoT appli- cations, such as connected vehicle and smart factory that require high availability, scalability, low latency, and security have appeared in the world. There have been many attempts to use cloud computing for IoT applications, but the mentioned requirements cannot be ensured in cloud environments. To solve this problem, edge computing has appeared in the world. In edge environments, containerization technology is useful to deploy apps with limited resources. In this thesis, two types of high available Kubernetes architecture (2 nodes with an external DB and 3 nodes with embedded DB) were surveyed and implemented using K3s distribution that is suitable for edges. By having a few experiments with the implemented K3s clusters, this thesis shows that the K3s clusters can provide high availability and scalability. We discuss the limitations of the implementations and provide possible solutions too. In addition, we provide the resource usages of each cluster in terms of CPU, RAM, and disk. Both clusters need only less than 10% CPU and about 500MB RAM on average. However, we could see that the 3 nodes cluster with embedded DB uses more resources than the 2 nodes + external DB cluster when changing the status of clusters. Finally, we show that the implemented K3s clusters are suitable for many IoT applications such as connected vehicle and smart factory. If an application that needs high availability and scalability has to be deployed in edge environments, the K3s clusters can provide good solutions to achieve the goals of the applications. The 2 nodes + external DB cluster is suitable for the applications where the amount of data fluctuate often, or where there is a stable connection with the external DB. On the other hand, the 3 nodes cluster will be suitable for the applications that need high availability of the database even in poor internet connection. ACM Computing Classification System (CCS) Computer systems organization → Embedded and cyber-physical systems Human-centered computing → Ubiquitous and mobile computing
  • Hyeongju, Lee (2021)
    The number of IoT and sensor devices is expected to reach 25 billion by 2030. Many IoT appli- cations, such as connected vehicle and smart factory that require high availability, scalability, low latency, and security have appeared in the world. There have been many attempts to use cloud computing for IoT applications, but the mentioned requirements cannot be ensured in cloud environments. To solve this problem, edge computing has appeared in the world. In edge environments, containerization technology is useful to deploy apps with limited resources. In this thesis, two types of high available Kubernetes architecture (2 nodes with an external DB and 3 nodes with embedded DB) were surveyed and implemented using K3s distribution that is suitable for edges. By having a few experiments with the implemented K3s clusters, this thesis shows that the K3s clusters can provide high availability and scalability. We discuss the limitations of the implementations and provide possible solutions too. In addition, we provide the resource usages of each cluster in terms of CPU, RAM, and disk. Both clusters need only less than 10% CPU and about 500MB RAM on average. However, we could see that the 3 nodes cluster with embedded DB uses more resources than the 2 nodes + external DB cluster when changing the status of clusters. Finally, we show that the implemented K3s clusters are suitable for many IoT applications such as connected vehicle and smart factory. If an application that needs high availability and scalability has to be deployed in edge environments, the K3s clusters can provide good solutions to achieve the goals of the applications. The 2 nodes + external DB cluster is suitable for the applications where the amount of data fluctuate often, or where there is a stable connection with the external DB. On the other hand, the 3 nodes cluster will be suitable for the applications that need high availability of the database even in poor internet connection. ACM Computing Classification System (CCS) Computer systems organization → Embedded and cyber-physical systems Human-centered computing → Ubiquitous and mobile computing
  • Sakaya, Joseph Hosanna (2015)
    Traditional natural language processing has been shown to have excessive reliance on human-annotated corpora. However, the recent successes of machine translation and speech recognition, ascribed to the effective use of the increasingly availability of web-scale data in the wild, has given momentum to a re-surging interest in attempting to model natural language with simple statistical models, such as the n-gram model, that are easily scaled. Indeed, words and word combinations provide all the representational machinery one needs for solving many natural language tasks. The degree of semantic similarity between two words is a function of the similarity of the linguistic contexts in which they appear. Word representations are mathematical objects, often vectors, that capture syntactic and semantic properties of a word. This results in words that are semantic cognates having similar word representations, an important property that we will widely use. We claim that word representations provide a superb framework for unsupervised learning on unlabelled data by compactly representing the distributional properties of words. The current state-of-the-art word representation adopts the skip-gram model to train shallow neural networks and presents negative sampling, an idea borrowed from Noise Contrastive Estimation, as an efficient method of inducing embeddings. An alternative approach contends that the inherent multi-contextual nature of words entails a more Canonical Correlation Analysis-like approach for best results. In this thesis we develop the first fully Bayesian model to induce word embeddings. The prominent contributions of this thesis are: 1. A crystallisation of the best practices from previous literature on word embeddings and matrix factorisation into a single hierarchical Bayesian model. 2. A scalable matrix factorisation technique for structured sparse data. 3. Representation of the latent dimensions as continuous Gaussian densities instead of as point estimates. We analyse a corpus of 170 million tokens and learn for each word form a vectorial representation based on the 8 surrounding context words with a negative sampling rate of 2 per token. We would like to stress that while we certainly hope to beat the state-of-the-art, our primary goal is to develop a stochastic and scalable Bayesian model. We evaluate the quality of the word embeddings against the word analogy tasks as well as other such tasks as word similarity and chunking. We demonstrate competitive performance on standard benchmarks.
  • Jiang, Tao (2017)
    Dually thermoresponsive poly(sulfobetaine methacylate)-graft-(poly(poly(ethylene glycol) methyl ether methacrylate)-co-poly(di(ethylene glycol) methyl ether methacrylate) were synthesized via single electron transfer living radical polymerization (SET-LRP). Two different such graft copolymers S70-g-P25D25 and S70-g-P70D280 with different side chain lengths were prepared and studied. These polymers showed 'schezophrenic' self-assembly behavior in response to temperature and ionic strength in aqeuous solution in water. S70-g-P25D25 formed nanostructures at temperatures both above the lower critical solution temperature (LCST) and below the upper critical solution temperature (UCST) with inversed core-shell nature in aqueous solution. Under saline condition no nano structure could be observed at temperatures below the UCST. For S70-g-P70D280, LCST type self-assembly was observed with the formation of similar nanostructures, but at temperatures below UCST, instead of intermolecular aggregation, unimolecular self-assembly was obsverved due to the much more crowded side chains.
  • Suomalainen, Aino (2020)
    This Master’s thesis studies the mechanisms connected to negative changes in educational outcomes in upper comprehensive schools in Helsinki. What are the factors associated with negative changes in educational outcomes of individual students during the transition from 7th to 9th grade? There is an increased socioeconomic and ethnic segregation in Helsinki Metropolitan Area, and the differences between schools’ levels of success have also been growing throughout the 21st century. There is little research on combining schools and city development in Finland. The aim is to examine is there an association between decreasing individual educational outcomes and socio-spatial or school segregation, and to look at what is the role of individual factors and social context in decreased educational outcomes. Studying pupils and schools is a good way to capture local processes of differentiation and neighbourhood effect, because children and youth are especially prone to neighbourhood and school effects due to their ongoing process of socialization, localized lives in their neighbourhood and shared institutions, such as school. This study is conducted quantitatively, and the main method in this study is hierarchical linear regression. The data is from Metropolitan Longitudinal Finland research, which studies the success and wellbeing of pupils in upper comprehensive schools in the Helsinki Metropolitan area. The study was conducted during the Fall of 2011 and the Spring of 2014 tracking the same cohort when the pupils were in their 7th and 9th grades. The results suggest that there are no differences found between schools, but some of the qualities describing neighborhoods indicate that some neighbourhood effect might be found. There are indications that pupils with decreased educational outcomes are more likely to study in schools that are located in low income areas than higher income areas. Also, for pupils with decreased educational outcomes, attending a school that is located in Northern or Southeastern Great districts is more likely than attending a school in Eastern Great district. Based on the results, pupils with negative change in educational outcomes are more likely to spend time with friends of own area than with school friends. Boys have a bigger risk for a negative change in educational outcomes than girls, and the change of school is connected to decreased educational outcomes. Mother’s education and immigration background was not found to have connection with decreased educational outcomes. Decreased educational outcomes have a connection with a low parents’ pedagogical ethos, but no connection with peers’ pedagogical ethos was found. The results are significant from the perspective of urban and educational politics and planning. The indications that the educational outcomes in upper comprehensive schools in Helsinki are differentiated in neighborhood level for example between Great districts, and in individual level between genders, challenge the goals of equal educational opportunities. Also, urban planning should be targeted to prevent socio-spatial differentiation of neighborhoods, in order to combat differentiation in schools’ composition of pupils. In future research, the starting level of educational success could be studied more closely- does decrease in educational outcomes implicate different educational paths for pupils that start with high starting level than pupils that have lower starting level in the beginning? This study provided information that there are no differences between schools found currently, but the processes of differentiation are not stable, so the processes should be observed continuously.
  • Jalkanen, Matias (2018)
    Tämän tutkielman tarkoitus on perehtyä analyyttisten funktioiden konformisuuteen ja erityisesti esittää Schawrz-Christoffelin kaava, joka on konformikuvaus ylemmältä puolitasolta yksinkertaiselle monikulmiolle. Tutkielma on jaettu kahteen osaan, joista ensimmäinen on Riemannin kuvauslause. Kappale on rakennettu niin, että perehdytään ensin analyyttisten funktioiden ominaisuuksiin, joita Riemannin kuvauslauseen todistuksessa tarvitaan. Näitä ovat logaritmin ominaisuudet, funktiojonon normaali suppeneminen ja konformisuus. Riemannin kuvauslauseen todistamisen jälkeen aloitetaan perehtymään Schwarz-Christoffelin kaavaan. Kappaleessa aloitetaan helposti ymmärrettävällä esimerkillä Schwarz-Christoffelin kaavan ideasta, jota kehitetään itse kaavaksi. Loppuosassa osoitetaan, että kaava on itseasiassa konforminen. Tämä vaatii tarkkaa analyysiä funktion käytöksestä kuvauksen reunalla.
  • Tenhunen, Jouni (2024)
    Compression methods are widely used in modern computing. With the amount of data stored and transferred by database systems constantly increasing, the implementation of compression methods into database systems has been studied from different angles during the past four decades. This thesis studies the scientific methods used in relational database research. The goal of the thesis is to evaluate the methods employed and to gain an understanding into how research into the subject should be conducted. A literature review is conducted. 14 papers are identified for review and their methodology is described and analysed. The papers reviewed are used to answer four research question and classified according to insights gained during the review process. There are similarities in methods of different papers that can be described to use as a starting point for conducting research in the field of database compression.
  • Romppainen, Mikko (Helsingin yliopistoHelsingfors universitetUniversity of Helsinki, 2010)
    Scrum-prosessimalli jättää menettelyt ohjelmiston käyttöliittymän tuottamiseen auki ja käyttöliittymä syntyy pahimmillaan toimintolistan pohjalta ohjelmoinnin sivutuotteena. Näin syntynyt järjestelmä soveltuu suurella riskillä huonosti käyttötarkoitukseensa. Tässä tutkielmassa analysoitiin Scrum-prosessimallin käyttöliittymäriskikohtia, joita löytyi kolme: asiakkaan toivomat ominaisuudet päätyvät sellaisinaan ohjelmiston vaatimuksiksi, toimintolistan pohjalta syntyy käytettävyydeltään heikkoja käyttöliittymäratkaisuja ja käyttöliittymän arviointi sprintin katselmoinnissa tuottaa epäluotettavia tuloksia. Tutkielmassa käsitellään Scrum-prosessimallin käyttöliittymäriskien minimointia simulointipohjaisella GDD-käyttöliittymäsuunnittelulla. Riskien minimointia tarkastellaan esimerkkitapauksen avulla, jossa ohjelmistoyritys Reaktor toteutti ammattikorkeakoulun toiminnansuunnittelujärjestelmän vuosisuunnitteluosion. Esimerkkitapauksessa Scrumin käyttöliittymäriskit saatiin minimoitua selvittämällä loppukäyttäjien käyttötilanteet kontekstuaalisilla haastatteluilla, suunnittelemalla käyttöliittymä GDD-menetelmällä ja arvioimalla käyttöliittymää hyödyllisyysläpikäynneillä. Alkuperäisessä Scrumissa liiketoiminnallisesta kannattavuudesta vastaava tuotteen omistaja ja toteutustiimi ottavat vastuulleen myös käyttöliittymän toimintalogiikan. GDD:n myötä vastuu toimintalogiikasta siirretään käyttöliittymäsuunnittelijalle, jolloin Scrumin roolit muuttuvat. Tässä työssä käsitellään GDD-käyttöliittymäsuunnittelun tuomat muutokset Scrumin rooleihin ja käytäntöihin. Scrumin käyttöliittymäriskien minimoinnin jälkeen toteutusvaiheeseen jää vielä Scrumista riippumattomia käyttöliittymäriskejä. Tämän työn esimerkkitapauksessa keskeisin näistä oli käyttöliittymätoteutukseen päätyneet puutteelliset interaktiot. Riski eliminoitiin hyväksymismenettelyllä, jossa ohjelmoija antaa toteutetun ominaisuuden käyttöliittymäsuunnittelijalle tarkistettavaksi. Hyväksymismenettelyn avulla projektin työnjako selkiytyi, toteutustyön laatu parani ja toteutustiimin ja käyttöliittymäsuunnittelijoiden välinen kommunikaatio tehostui.
  • Virta, Maxim (2022)
    Strongly coupled matter called quark–gluon plasma (QGP) is formed in heavy-ion collisions at RHIC [1, 2] and the LHC [3, 4]. The expansion of this matter, caused by pressure gradients, is known to be hydrodynamic expansion. The computations show that the expanding QGP has a small shear viscosity to entropy density ratio (η/s), close to the known lower bound 1/4π [5]. In such a medium one expects that jets passing through the medium would create Mach cones. Many experimental trials have been done but no proper evidence of the Mach cone has been found [6, 7]. Mach cones were thought to cause double-bumps in azimuthal correlations. However these were later shown to be the third flow harmonic. In this thesis a new method is proposed for finding the Mach cone with so called event-engineering. The higher order flow harmonics and their linear response are known to be sensitive to the medium properties [8]. Hence a Mach cone produced by high momentum jet would change the system properties and, thus, the observable yields. Different flow observables are then studied by selecting high energy jet events with different momentum ranges. Considered observables for different momenta are then compared to the all events. Found differences in the flow harmonics and their correlations for different jet momenta are reported showing evidence of Mach cone formation in the heavy-ion collisions. The observations for different jet momenta are then quantified with χ 2 -test to see the sensitivities of different observables to momentum selections.
  • Heikkilä, Jaana Kristiina (2015)
    ATLAS and CMS collaborations at the Large Hadron Collider have observed a new resonance consistent with the standard model Higgs boson. However, it has been suggested that the observed signal could also be produced by multiple nearly mass-degenerate states that couple differently to the standard model particles. In this work, a method to discriminate between the hypothesis of a single Higgs boson and that of multiple mass-degenerate Higgs bosons was developed. Using the matrix of measured signal strengths in different production and decay modes, parametrizations for the two hypotheses were constructed as a general rank 1 matrix and the most general 5 x 4 matrix, respectively. The test statistic was defined as a ratio of profile likelihoods for the two hypotheses. The method was applied to the CMS measurements. The expected test statistic distribution was estimated twice by generating pseudo-experiments according to both the standard model hypothesis and the single Higgs boson hypothesis best fitting the data. The p-value for the single Higgs boson hypothesis was defined from both expected test statistic distributions, and it was (8.0 ± 0.3)% and (11.0±0.3)%, respectively. In addition to this, a p-value was also estimated in an alternative way using a χ2 distribution, fitted to the pseudo-experiments for the standard model Higgs boson hypothesis. The resulting p-value was 10.8%. Thus the three estimates yield similar p-values for the single Higgs boson hypothesis. These results suggest that the CMS data is compatible with the single Higgs boson hypothesis, as in the standard model. Furthermore, the result is insensitive to choice of the single Higgs boson hypothesis used to derive it, and it does not depend on the precise shape of the test statistic distribution. The developed method can be applied also to other arbitrarily-sized matrices, and it takes into account the uncertainties on the measurements, missing elements of data, and possible correlations. This thesis is an extensive description of the method that has also been published in EPJC (David, Heikkilä and Petrucciani), and the method has been used in the final Run 1 Higgs combination and properties article (CMS Collaboration, incl. Heikkilä).
  • Huhtilainen, Heli (2023)
    This thesis studies the application of language models to improve search in an online shop specialising in wholesale builder–trade product and service sales. The first aim was to determine if a Finnish language model could capture the meaning behind the search query words to improve the match between the queries and the product descriptions. Secondly, it was investigated if it was possible to train the model to recognise what products the users wanted to find with the search terms they used. Finally, it was investigated if it was possible to use the model for search ranking. Three models were trained using FinBERT as a model checkpoint and domain-specific product and clickthrough data for fine-tuning the models. The first two models were trained to classify online store products into product categories. The task was completed with a 0.98 F measure score for the model with 55 target categories and a 0.81 F measure score for the model with 762 target categories. The third model was trained to evaluate the relevance probabilities of search query-product pairs. The model generally determined more products as relevant than the current search engine solution. The F measure score for the model was 0.90, and in qualitative evaluation, the predictions made by the model made semantically sense. The restrictions for the practical use of the third model for search ranking come from the prediction inference needing to be faster to make search ranking predictions for many products.
  • Uusinoka, Matias (2022)
    Sea-ice dynamics is becoming increasingly essential for the modelling warming climate as the extent and thickness of the ice cover are decreasing along with increasing drift speeds and mechanical weakening. The description of the sea-ice dynamics involves an enormous variety of spatial and temporal scales from meters to the scale of the Arctic Basin and from seconds to years in the geophysical approaches. The complex coupled spatio-temporal scaling laws prohibit the commonly utilized procedures for scale linkage of ice mechanics. Currently, deformation scaling presents one of the principal open questions in sea ice dynamics for which the thesis aims to provide observational analysis. The high-resolution ship-radar imagery gathered during the MOSAiC expedition from October 2019 to September 2020 for which deformation component rates were calculated to generate a seasonal deformation time series. Current research of deformation scaling commonly relies on satellite imagery and drift buoys for which the spatial and temporal resolutions often tend to be considerably lower than for the ship-radar data. The formerly observed dominant deformation mode of shear and the strong spatial correlation of divergence and shear in the Arctic sea ice were confirmed with no signs of seasonal variation. The temporally averaged deformation variations were found to coincide with satellite derived deformation events rather poorly. A strong length scale dependence of deformation was confirmed in the ship-radar data. The spatial scaling law exponents were found to show unexpectedly high values with the behaviour of both spatial and temporal scaling law exponents disobeying the previously observed large-scale characteristics. The seasonal variation of both scaling law exponents were found to exhibit the commonly observed trends following the progression of total deformation rate. The obtained results showed unexpected values and behaviour for the deformation scaling law exponents, which was suggested to be due to the technical faults in the ship-radar data. The faults were often spatially local and lasted merely for a single time step leading to a possible increase in the localization and intermittency of the deformation rates. Additionally, the new ice conditions of the Arctic Ocean and drift route along the Transpolar Drift were suggested as a possible physical source of the unexpected results. Further studies with different methodologies were suggested for the verification and possible the dismissal of the unexpected results.
  • Bloss, Matthew (2016)
    Atmospheric secondary organic aerosol formation is not completely understood, aerosol processes and chemistry are complex as well as their interactions and evolution throughout the atmosphere. In this study the potential aerosol mass chamber was used to simulate these processes. Potential aerosol mass being the maximum mass that can be produced from precursor gases being oxidised. UV lamps situated inside the chamber produce oxidants (O3, OH, HO2) at the same ratio as the atmosphere and the lamp intensity can be changed to simulate different days of photochemical aging. Laboratory characterisation measurements were undertaken by adjusting various experimental parameters (level of oxidants, relative humidity, temperature, type of precursor and its concentration). The aim of these measurements is to understand how the PAM chamber simulates atmospheric aerosol aging and photochemical processes. The PAM chamber was observed to simulate atmospheric SOA formation well. As the amount of oxidants was increased the amount of formed SOA increased initially. After a certain point, it was observed that fragmentation started to decrease the mass even though the oxidation level of SOA increased further.
  • Ding, Yi (Helsingin yliopistoHelsingfors universitetUniversity of Helsinki, 2009)
    In wireless and mobile networks, handover is a complex process that involves multiple layers of protocol and security executions. With the growing popularity of real time communication services such as Voice of IP, a great challenge faced by handover nowadays comes from the impact of security implementations that can cause performance degradation especially for mobile devices with limited resources. Given the existing networks with heterogeneous wireless access technologies, one essential research question that needs be addressed is how to achieve a balance between security and performance during the handover. The variations of security policy and agreement among different services and network vendors make the topic challenging even more, due to the involvement of commercial and social factors. In order to understand the problems and challenges in this field, we study the properties of handover as well as state of the art security schemes to assist handover in wireless IP networks. Based on our analysis, we define a two-phase model to identify the key procedures of handover security in wireless and mobile networks. Through the model we analyze the performance impact from existing security schemes in terms of handover completion time, throughput, and Quality of Services (QoS). As our endeavor of seeking a balance between handover security and performance, we propose the local administrative domain as a security enhanced localized domain to promote the handover performance. To evaluate the performance improvement in local administrative domain, we implement the security protocols adopted by our proposal in the ns-2 simulation environment and analyze the measurement results based on our simulation test.
  • Atarah, Ivan Akoribila (2017)
    Connected devices of IoT platforms are known to produce, process and exchange vast amounts of data, most of it sensitive or personal, that need to be protected. However, achieving minimal data protection requirements such as confidentiality, integrity, availability and non-repudiation in IoT platforms is a non-trivial issue. For one reason, the trillions of interacting devices provide larger attack surfaces. Secondly, high levels of personal and private data sharing in this ubiquitous and heterogeneous environment require more stringent protection. Additionally, whilst interoperability fuels innovation through cross-platform data flow, data ownership is a concern. This calls for categorizing data and providing different levels of access control to users known as global and local scopes. These issues present new and unique security considerations in IoT products and services that need to be addressed to enable wide adoption of the IoT paradigm. This thesis presents a security and privacy framework for the Web of IoT platforms that addresses end-to-end security and privacy needs of the platforms. It categorizes platforms’ resources into different levels of security requirements and provides appropriate access control mechanisms.
  • Zhang, Yu (2022)
    The Internet of Things (IoT) aims at linking billions of devices using the internet and other heterogeneous networks to share information. However, the issues of security in IoT environments are more challenging than with ordinary Internet. A vast number of devices are exposed to the attackers, and some of those devices contain sensitive personal and confidential data. For example, the sensitive flows of data such as autonomous vehicles, patient life support devices, traffic data in smart cities are extremely concerned by researchers from the security field. The IoT architecture needs to handle security and privacy requirements such as provision of authentication, access control, privacy and confidentiality. This thesis presents the architecture of IoT and its security issues. Additionally, we introduce the concept of blockchain technology, and the role of blockchain in different security aspects of IoT is discussed through a literature review. In case study of Mirai, we explain how snort and iptables based approach can be used to prevent IoT botnet from finding IoT devices by port scanning.
  • Kukkonen, Tommi (2022)
    Eutrophication and harmful substances of anthropogenic origin threaten the state of the Baltic Sea and especially its geochemistry and oxygen levels near the seafloor. Water exchange between the Baltic Sea and the Atlantic Ocean can affect oxygen circulation and sedimentation rates, but they are considered very sporadic and it is unclear how the water circulation and flow rates affect element concentrations and sediment deposition in the near seafloor environments. One of the less studied basins is the Western Gulf of Finland and its seafloor environment. During the 2019 voyage, the seafloor located to the south of the city of Hanko was investigated through bathymetric sounding tools and other measurements in which element concentration and sediment deposition rate data was acquired. The sounding revealed a large channel cutting the seafloor which was hypothesized to influence the nearbottom conditions. The obtained data consisted of samples from 13 short, 40 cm sediment cores which were analysed for 137Cs activity, organic content, and grain size distribution. The goal of the thesis was to determine the intensity of water exchange taking place in the seabed channels between the mid-Baltic Sea and the Western Gulf of Finland and investigate the effect of the seafloor channel and flow rates on sediment and element deposition, their relationships, and how they affect the overall conditions in the study area. These relationships were analyzed through spatial and statistical methods by utilizing GIS-tools to interpolate the data obtained from the study locations by using the Inverse Distance Weighting (IDW) method, and by multielement analyses in the R-environment, namely Principal Component Analysis (PCA) and Partial Least Regression (PLS) to analyze grain size and element concentration correlations and combine them with obtained flow rate data. The results showed strong correlation in flow rate intensities between the Western Gulf of Finland and the mid-Baltic Sea, and they are strongly linked with sedimentation and element deposition rates. However, no long-term trend was identified for the seafloor channel velocity frequencies. The Cs-activity shows stronger sedimentation activity on the western side of the seafloor channel. The overall element and sediment deposition in the study area was largely controlled by monthly and seasonal current velocity fluctuations among other processes. The element concentration comparison showed weakened oxygen conditions in the study area with increased eutrophication and carbon burial since the 1950s. The principal Component Analysis showed smaller grain sizes (0.15 - 2 mm) having a stronger influence on the datasets with Mo, N, and C providing largest variation in the data. Interpolation showed oxygen, pH, and H2S to have more fluctuation in the study area, which can indicate changes in the vertical gradients in each sample point. It could also be determined that other measured concentrations, such as temperature, turbidity, and salinity do not respond very sensitively to water inflow fluctuations or sedimentation rate changes. The results indicate that harmful substances and eutrophication are most likely going to increase in the near-bottom environment in the Western Gulf of Finland, contributed by anthropogenic activity. Water exchange is likely to become more and more uneven, thus affecting the flow rate effects to the sediment deposition in the Baltic Sea. Further studies are needed to link these processes to large-scale global changes and the general state of changes happening in the Baltic Sea and its surrounding areas. The seafloor of the Western Gulf of Finland could also be studied further to gain a better understanding of longer timescale changes on the seafloor channel currents, and element and sediment deposition rates.