machine learning thesis topics

Analytics Insight

Top 10 Research and Thesis Topics for ML Projects in 2022

Avatar photo

This article features the top 10 research and thesis topics for ML projects for students to try in 2022

Text mining and text classification, image-based applications, machine vision, optimization, voice classification, sentiment analysis, recommendation framework project, mall customers’ project, object detection with deep learning.

Whatsapp Icon

Disclaimer: Any financial and crypto market information given on Analytics Insight are sponsored articles, written for informational purpose only and is not an investment advice. The readers are further advised that Crypto products and NFTs are unregulated and can be highly risky. There may be no regulatory recourse for any loss from such transactions. Conduct your own research by contacting financial experts before making any investment decisions. The decision to read hereinafter is purely a matter of choice and shall be construed as an express undertaking/guarantee in favour of Analytics Insight of being absolved from any/ all potential legal action, or enforceable claims. We do not represent nor own any cryptocurrency, any complaints, abuse or concerns with regards to the information provided shall be immediately informed here .

You May Also Like

Bitcoin

Amid Bitcoin Ordinals Hype, Bitcoin Frogs Become Most Traded NFT

Dogetti

Why Crypto Surge Following SVB Collapse? Avoid FOMO with Shiba Inu, Dogecoin, and Dogetti

Bitcoin

Bitcoin Price Prediction: Can BTC Hit $40,000 in 2023?

Budget 2024

Budget 2024: Education Sector Seeks VR, AI, and Funding Support

machine learning thesis topics

Analytics Insight® is an influential platform dedicated to insights, trends, and opinion from the world of data-driven technologies. It monitors developments, recognition, and achievements made by Artificial Intelligence, Big Data and Analytics companies across the globe.

linkedin

  • Select Language:
  • Privacy Policy
  • Content Licensing
  • Terms & Conditions
  • Submit an Interview

Special Editions

  • Dec – Crypto Weekly Vol-1
  • 40 Under 40 Innovators
  • Women In Technology
  • Market Reports
  • AI Glossary
  • Infographics

Latest Issue

Magazine Issue January 2024

Disclaimer: Any financial and crypto market information given on Analytics Insight is written for informational purpose only and is not an investment advice. Conduct your own research by contacting financial experts before making any investment decisions, more information here .

Second Menu

Also, note that the cryptocurrencies mentioned/listed on the website could potentially be scams. i.e designed to induce you to invest financial resources that may be lost forever and not be recoverable once investments are made. You are resposible for conducting your ownresearch (DYOR) before making any investment.

Available Master's thesis topics in machine learning

Main content.

Here we list topics that are available. You may also be interested in our list of completed Master's theses .

Learning and inference with large Bayesian networks

Most learning and inference tasks with Bayesian networks are NP-hard. Therefore, one often resorts to using different heuristics that do not give any quality guarantees.

Task: Evaluate quality of large-scale learning or inference algorithms empirically.

Advisor: Pekka Parviainen

Sum-product networks

Traditionally, probabilistic graphical models use a graph structure to represent dependencies and independencies between random variables. Sum-product networks are a relatively new type of a graphical model where the graphical structure models computations and not the relationships between variables. The benefit of this representation is that inference (computing conditional probabilities) can be done in linear time with respect to the size of the network.

Potential thesis topics in this area: a) Compare inference speed with sum-product networks and Bayesian networks. Characterize situations when one model is better than the other. b) Learning the sum-product networks is done using heuristic algorithms. What is the effect of approximation in practice?

Bayesian Bayesian networks

The naming of Bayesian networks is somewhat misleading because there is nothing Bayesian in them per se; A Bayesian network is just a representation of a joint probability distribution. One can, of course, use a Bayesian network while doing Bayesian inference. One can also learn Bayesian networks in a Bayesian way. That is, instead of finding an optimal network one computes the posterior distribution over networks.

Task: Develop algorithms for Bayesian learning of Bayesian networks (e.g., MCMC, variational inference, EM)

Large-scale (probabilistic) matrix factorization

The idea behind matrix factorization is to represent a large data matrix as a product of two or more smaller matrices.They are often used in, for example, dimensionality reduction and recommendation systems. Probabilistic matrix factorization methods can be used to quantify uncertainty in recommendations. However, large-scale (probabilistic) matrix factorization is computationally challenging.

Potential thesis topics in this area: a) Develop scalable methods for large-scale matrix factorization (non-probabilistic or probabilistic), b) Develop probabilistic methods for implicit feedback (e.g., recommmendation engine when there are no rankings but only knowledge whether a customer has bought an item)

Bayesian deep learning

Standard deep neural networks do not quantify uncertainty in predictions. On the other hand, Bayesian methods provide a principled way to handle uncertainty. Combining these approaches leads to Bayesian neural networks. The challenge is that Bayesian neural networks can be cumbersome to use and difficult to learn.

The task is to analyze Bayesian neural networks and different inference algorithms in some simple setting.

Deep learning for combinatorial problems

Deep learning is usually applied in regression or classification problems. However, there has been some recent work on using deep learning to develop heuristics for combinatorial optimization problems; see, e.g., [1] and [2].

Task: Choose a combinatorial problem (or several related problems) and develop deep learning methods to solve them.

References: [1] Vinyals, Fortunato and Jaitly: Pointer networks. NIPS 2015. [2] Dai, Khalil, Zhang, Dilkina and Song: Learning Combinatorial Optimization Algorithms over Graphs. NIPS 2017.

Advisors: Pekka Parviainen, Ahmad Hemmati

Estimating the number of modes of an unknown function

Mode seeking considers estimating the number of local maxima of a function f. Sometimes one can find modes by, e.g., looking for points where the derivative of the function is zero. However, often the function is unknown and we have only access to some (possibly noisy) values of the function. 

In topological data analysis,  we can analyze topological structures using persistent homologies. For 1-dimensional signals, this can translate into looking at the birth/death persistence diagram, i.e. the birth and death of connected topological components as we expand the space around each point where we have observed our function. These observations turn out to be closely related to the modes (local maxima) of the function. A recent paper [1] proposed an efficient method for mode seeking.

In this project, the task is to extend the ideas from [1] to get a probabilistic estimate on the number of modes. To this end, one has to use probabilistic methods such as Gaussian processes.

[1] U. Bauer, A. Munk, H. Sieling, and M. Wardetzky. Persistence barcodes versus Kolmogorov signatures: Detecting modes of one-dimensional signals. Foundations of computational mathematics17:1 - 33, 2017.

Advisors:  Pekka Parviainen ,  Nello Blaser

Causal Abstraction Learning

We naturally make sense of the world around us by working out causal relationships between objects and by representing in our minds these objects with different degrees of approximation and detail. Both processes are essential to our understanding of reality, and likely to be fundamental for developing artificial intelligence. The first process may be expressed using the formalism of structural causal models, while the second can be grounded in the theory of causal abstraction.        

This project will consider the problem of learning an abstraction between two given structural causal models. The primary goal will be the development of efficient algorithms able to learn a meaningful abstraction between the given causal models.

Advisor: Fabio Massimo Zennaro

Causal Bandits

"Multi-armed bandit" is an informal name for slot machines, and the formal name of a large class of problems where an agent has to choose an action among a range of possibilities without knowing the ensuing rewards. Multi-armed bandit problems are one of the most essential reinforcement learning problems where an agent is directly faced with an exploitation-exploration trade-off.

This project will consider a class of multi-armed bandits where an agent, upon taking an action, interacts with a causal system. The primary goal will be the development of learning strategies that takes advantage of the underlying causal system in order to learn optimal policies in a shortest amount of time.

Causal Modelling for Battery Manufacturing

Lithium-ion batteries are poised to be one of the most important sources of energy in the near future. Yet, the process of manufacturing these batteries is very hard to model and control. Optimizing the different phases of production to maximize the lifetime of the batteries is a non-trivial challenge since physical models are limited in scope and collecting experimental data is extremely expensive and time-consuming.        

This project will consider the problem of aggregating and analyzing data regarding a few stages in the process of battery manufacturing. The primary goal will be the development of algorithms for transporting and integrating data collected in different contexts, as well as the use of explainable algorithms to interpret them.

Reinforcement Learning for Computer Security

The field of computer security presents a wide variety of challenging problems for artificial intelligence and autonomous agents. Guaranteeing the security of a system against attacks and penetrations by malicious hackers has always been a central concern of this field, and machine learning could now offer a substantial contribution. Security capture-the-flag simulations are particularly well-suited as a testbed for the application and development of reinforcement learning algorithms.

This project will consider the use of reinforcement learning for the preventive purpose of testing systems and discovering vulnerabilities before they can be exploited. The primary goal will be the modelling of capture-the-flag challenges of interest and the development of reinforcement learning algorithms that can solve them.

Approaches to AI Safety

The world and the Internet are more and more populated by artificial autonomous agents carrying out tasks on our behalf. Many of these agents are provided with an objective and they learn their behaviour trying to achieve their objective as best as they can. However, this approach can not guarantee that an agent, while learning its behaviour, will not undertake actions that may have unforeseen and undesirable effects. Research in AI safety tries to design autonomous agent that will behave in a predictable and safe way. 

This project will consider specific problems and novel solution in the domain of AI safety and reinforcement learning. The primary goal will be the development of innovative algorithms and their implementation withing established frameworks.

The Topology of Flight Paths

Air traffic data tells us the position, direction, and speed of an aircraft at a given time. In other words, if we restrict our focus to a single aircraft, we are looking at a multivariate time-series. We can visualize the flight path as a curve above earth's surface quite geometrically. Topological data analysis (TDA) provides different methods for analysing the shape of data. Consequently, TDA may help us to extract meaningful features from the air traffic data. Although the typical flight path shapes may not be particularly intriguing, we can attempt to identify more intriguing patterns or “abnormal” manoeuvres, such as aborted landings, go-arounds, or diverts.

Advisor:  Odin Hoff Gardå , Nello Blaser

Automatic hyperparameter selection for isomap

Isomap is a non-linear dimensionality reduction method with two free hyperparameters (number of nearest neighbors and neighborhood radius). Different hyperparameters result in dramatically different embeddings. Previous methods for selecting hyperparameters focused on choosing one optimal hyperparameter. In this project, you will explore the use of persistent homology to find parameter ranges that result in stable embeddings. The project has theoretic and computational aspects.

Advisor: Nello Blaser

Validate persistent homology

Persistent homology is a generalization of hierarchical clustering to find more structure than just the clusters. Traditionally, hierarchical clustering has been evaluated using resampling methods and assessing stability properties. In this project you will generalize these resampling methods to develop novel stability properties that can be used to assess persistent homology. This project has theoretic and computational aspects.

Topological Ancombs quartet

This topic is based on the classical Ancombs quartet and families of point sets with identical 1D persistence ( https://arxiv.org/abs/2202.00577 ). The goal is to generate more interesting datasets using the simulated annealing methods presented in ( http://library.usc.edu.ph/ACM/CHI%202017/1proc/p1290.pdf ). This project is mostly computational.

Persistent homology vectorization with cycle location

There are many methods of vectorizing persistence diagrams, such as persistence landscapes, persistence images, PersLay and statistical summaries. Recently we have designed algorithms to in some cases efficiently detect the location of persistence cycles. In this project, you will vectorize not just the persistence diagram, but additional information such as the location of these cycles. This project is mostly computational with some theoretic aspects.

Divisive covers

Divisive covers are a divisive technique for generating filtered simplicial complexes. They original used a naive way of dividing data into a cover. In this project, you will explore different methods of dividing space, based on principle component analysis, support vector machines and k-means clustering. In addition, you will explore methods of using divisive covers for classification. This project will be mostly computational.

Learning Acquisition Functions for Cost-aware Bayesian Optimization

This is a follow-up project of an earlier Master thesis that developed a novel method for learning Acquisition Functions in Bayesian Optimization through the use of Reinforcement Learning. The goal of this project is to further generalize this method (more general input, learned cost-functions) and apply it to hyperparameter optimization for neural networks.

Advisors: Nello Blaser , Audun Ljone Henriksen

Stable updates

This is a follow-up project of an earlier Master thesis that introduced and studied empirical stability in the context of tree-based models. The goal of this project is to develop stable update methods for deep learning models. You will design sevaral stable methods and empirically compare them (in terms of loss and stability) with a baseline and with one another.

Advisors:  Morten Blørstad , Nello Blaser

Multimodality in Bayesian neural network ensembles

One method to assess uncertainty in neural network predictions is to use dropout or noise generators at prediction time and run every prediction many times. This leads to a distribution of predictions. Informatively summarizing such probability distributions is a non-trivial task and the commonly used means and standard deviations result in the loss of crucial information, especially in the case of multimodal distributions with distinct likely outcomes. In this project, you will analyze such multimodal distributions with mixture models and develop ways to exploit such multimodality to improve training. This project can have theoretical, computational and applied aspects.

Learning a hierarchical metric

Often, labels have defined relationships to each other, for instance in a hierarchical taxonomy. E.g. ImageNet labels are derived from the WordNet graph, and biological species are taxonomically related, and can have similarities depending on life stage, sex, or other properties.

ArcFace is an alternative loss function that aims for an embedding that is more generally useful than softmax. It is commonly used in metric learning/few shot learning cases.

Here, we will develop a metric learning method that learns from data with hierarchical labels. Using multiple ArcFace heads, we will simultaneously learn to place representations to optimize the leaf label as well as intermediate labels on the path from leaf to root of the label tree. Using taxonomically classified plankton image data, we will measure performance as a function of ArcFace parameters (sharpness/temperature and margins -- class-wise or level-wise), and compare the results to existing methods.

Advisor: Ketil Malde ( [email protected] )

Self-supervised object detection in video

One challenge with learning object detection is that in many scenes that stretch off into the distance, annotating small, far-off, or blurred objects is difficult. It is therefore desirable to learn from incompletely annotated scenes, and one-shot object detectors may suffer from incompletely annotated training data.

To address this, we will use a region-propsal algorithm (e.g. SelectiveSearch) to extract potential crops from each frame. Classification will be based on two approaches: a) training based on annotated fish vs random similarly-sized crops without annotations, and b) using a self-supervised method to build a representation for crops, and building a classifier for the extracted regions. The method will be evaluated against one-shot detectors and other training regimes.

If successful, the method will be applied to fish detection and tracking in videos from baited and unbaited underwater traps, and used to estimate abundance of various fish species.

See also: Benettino (2016): https://link.springer.com/chapter/10.1007/978-3-319-48881-3_56

Representation learning for object detection

While traditional classifiers work well with data that is labeled with disjoint classes and reasonably balanced class abundances, reality is often less clean. An alternative is to learn a vectors space embedding that reflects semantic relationships between objects, and deriving classes from this representation. This is especially useful for few-shot classification (ie. very few examples in the training data).

The task here is to extend a modern object detector (e.g. Yolo v8) to output an embedding of the identified object. Instead of a softmax classifier, we can learn the embedding either in a supervised manner (using annotations on frames) by attaching an ArcFace or other supervised metric learning head. Alternatively, the representation can be learned from tracked detections over time using e.g. a contrastive loss function to keep the representation for an object (approximately) constant over time. The performance of the resulting object detector will be measured on underwater videos, targeting species detection and/or indiviual recognition (re-ID).

Time-domain object detection

Object detectors for video are normally trained on still frames, but it is evident (from human experience) that using time domain information is more effective. I.e., it can be hard to identify far-off or occluded objects in still images, but movement in time often reveals them.

Here we will extend a state of the art object detector (e.g. yolo v8) with time domain data. Instead of using a single frame as input, the model will be modified to take a set of frames surrounding the annotated frame as input. Performance will be compared to using single-frame detection.

Large-scale visualization of acoustic data

The Institute of Marine Research has decades of acoustic data collected in various surveys. These data are in the process of being converted to data formats that can be processed and analyzed more easily using packages like Xarray and Dask.

The objective is to make these data more accessible to regular users by providing a visual front end. The user should be able to quickly zoom in and out, perform selection, export subsets, apply various filters and classifiers, and overlay annotations and other relevant auxiliary data.

Learning acoustic target classification from simulation

Broadband echosounders emit a complex signal that spans a large frequency band. Different targets will reflect, absorb, and generate resonance at different amplitudes and frequencies, and it is therefore possible to classify targets at much higher resolution and accuracy than before. Due to the complexity of the received signals, deriving effective profiles that can be used to identify targets is difficult.

Here we will use simulated frequency spectra from geometric objects with various shapes, orientation, and other properties. We will train ML models to estimate (recover) the geometric and material properties of objects based on these spectra. The resulting model will be applied to read broadband data, and compared to traditional classification methods.

Online learning in real-time systems

Build a model for the drilling process by using the Virtual simulator OpenLab ( https://openlab.app/ ) for real-time data generation and online learning techniques. The student will also do a short survey of existing online learning techniques and learn how to cope with errors and delays in the data.

Advisor: Rodica Mihai

Building a finite state automaton for the drilling process by using queries and counterexamples

Datasets will be generated by using the Virtual simulator OpenLab ( https://openlab.app/ ). The student will study the datasets and decide upon a good setting to extract a finite state automaton for the drilling process. The student will also do a short survey of existing techniques for extracting finite state automata from process data. We present a novel algorithm that uses exact learning and abstraction to extract a deterministic finite automaton describing the state dynamics of a given trained RNN. We do this using Angluin's L*algorithm as a learner and the trained RNN as an oracle. Our technique efficiently extracts accurate automata from trained RNNs, even when the state vectors are large and require fine differentiation.arxiv.org

Scaling Laws for Language Models in Generative AI

Large Language Models (LLM) power today's most prominent language technologies in Generative AI like ChatGPT, which, in turn, are changing the way that people access information and solve tasks of many kinds.

A recent interest on scaling laws for LLMs has shown trends on understanding how well they perform in terms of factors like the how much training data is used, how powerful the models are, or how much computational cost is allocated. (See, for example, Kaplan et al. - "Scaling Laws for Neural Language Models”, 2020.)

In this project, the task will consider to study scaling laws for different language models and with respect with one or multiple modeling factors.

Advisor: Dario Garigliotti

Applications of causal inference methods to omics data

Many hard problems in machine learning are directly linked to causality [1]. The graphical causal inference framework developed by Judea Pearl can be traced back to pioneering work by Sewall Wright on path analysis in genetics and has inspired research in artificial intelligence (AI) [1].

The Michoel group has developed the open-source tool Findr [2] which provides efficient implementations of mediation and instrumental variable methods for applications to large sets of omics data (genomics, transcriptomics, etc.). Findr works well on a recent data set for yeast [3].

We encourage students to explore promising connections between the fiels of causal inference and machine learning. Feel free to contact us to discuss projects related to causal inference. Possible topics include: a) improving methods based on structural causal models, b) evaluating causal inference methods on data for model organisms, c) comparing methods based on causal models and neural network approaches.

References:

1. Schölkopf B, Causality for Machine Learning, arXiv (2019):  https://arxiv.org/abs/1911.10500

2. Wang L and Michoel T. Efficient and accurate causal inference with hidden confounders from genome-transcriptome variation data. PLoS Computational Biology 13:e1005703 (2017).  https://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1005703

3. Ludl A and and Michoel T. Comparison between instrumental variable and mediation-based methods for reconstructing causal gene networks in yeast. arXiv:2010.07417  https://arxiv.org/abs/2010.07417

Advisors: Adriaan Ludl ,  Tom Michoel

Space-Time Linkage of Fish Distribution to Environmental Conditions

Conditions in the marine environment, such as, temperature and currents, influence the spatial distribution and migration patterns of marine species. Hence, understanding the link between environmental factors and fish behavior is crucial in predicting, e.g., how fish populations may respond to climate change.   Deriving this link is challenging because it requires analysis of two types of datasets (i) large environmental (currents, temperature) datasets that vary in space and time, and (ii) sparse and sporadic spatial observations of fish populations.

Project goal   

The primary goal of the project is to develop a methodology that helps predict how spatial distribution of two fish stocks (capelin and mackerel) change in response to variability in the physical marine environment (ocean currents and temperature).  The information can also be used to optimize data collection by minimizing time spent in spatial sampling of the populations.

The project will focus on the use of machine learning and/or causal inference algorithms.  As a first step, we use synthetic (fish and environmental) data from analytic models that couple the two data sources.  Because the ‘truth’ is known, we can judge the efficiency and error margins of the methodologies. We then apply the methodologies to real world (empirical) observations.

Advisors:  Tom Michoel , Sam Subbey . 

Towards precision medicine for cancer patient stratification

On average, a drug or a treatment is effective in only about half of patients who take it. This means patients need to try several until they find one that is effective at the cost of side effects associated with every treatment. The ultimate goal of precision medicine is to provide a treatment best suited for every individual. Sequencing technologies have now made genomics data available in abundance to be used towards this goal.

In this project we will specifically focus on cancer. Most cancer patients get a particular treatment based on the cancer type and the stage, though different individuals will react differently to a treatment. It is now well established that genetic mutations cause cancer growth and spreading and importantly, these mutations are different in individual patients. The aim of this project is use genomic data allow to better stratification of cancer patients, to predict the treatment most likely to work. Specifically, the project will use machine learning approach to integrate genomic data and build a classifier for stratification of cancer patients.

Advisor: Anagha Joshi

Unraveling gene regulation from single cell data

Multi-cellularity is achieved by precise control of gene expression during development and differentiation and aberrations of this process leads to disease. A key regulatory process in gene regulation is at the transcriptional level where epigenetic and transcriptional regulators control the spatial and temporal expression of the target genes in response to environmental, developmental, and physiological cues obtained from a signalling cascade. The rapid advances in sequencing technology has now made it feasible to study this process by understanding the genomewide patterns of diverse epigenetic and transcription factors as well as at a single cell level.

Single cell RNA sequencing is highly important, particularly in cancer as it allows exploration of heterogenous tumor sample, obstructing therapeutic targeting which leads to poor survival. Despite huge clinical relevance and potential, analysis of single cell RNA-seq data is challenging. In this project, we will develop strategies to infer gene regulatory networks using network inference approaches (both supervised and un-supervised). It will be primarily tested on the single cell datasets in the context of cancer.

Developing a Stress Granule Classifier

To carry out the multitude of functions 'expected' from a human cell, the cell employs a strategy of division of labour, whereby sub-cellular organelles carry out distinct functions. Thus we traditionally understand organelles as distinct units defined both functionally and physically with a distinct shape and size range. More recently a new class of organelles have been discovered that are assembled and dissolved on demand and are composed of liquid droplets or 'granules'. Granules show many properties characteristic of liquids, such as flow and wetting, but they can also assume many shapes and indeed also fluctuate in shape. One such liquid organelle is a stress granule (SG). 

Stress granules are pro-survival organelles that assemble in response to cellular stress and important in cancer and neurodegenerative diseases like Alzheimer's. They are liquid or gel-like and can assume varying sizes and shapes depending on their cellular composition. 

In a given experiment we are able to image the entire cell over a time series of 1000 frames; from which we extract a rough estimation of the size and shape of each granule. Our current method is susceptible to noise and a granule may be falsely rejected if the boundary is drawn poorly in a small majority of frames. Ideally, we would also like to identify potentially interesting features, such as voids, in the accepted granules.

We are interested in applying a machine learning approach to develop a descriptor for a 'classic' granule and furthermore classify them into different functional groups based on disease status of the cell. This method would be applied across thousands of granules imaged from control and disease cells. We are a multi-disciplinary group consisting of biologists, computational scientists and physicists. 

Advisors: Sushma Grellscheid , Carl Jones

Machine Learning based Hyperheuristic algorithm

Develop a Machine Learning based Hyper-heuristic algorithm to solve a pickup and delivery problem. A hyper-heuristic is a heuristics that choose heuristics automatically. Hyper-heuristic seeks to automate the process of selecting, combining, generating or adapting several simpler heuristics to efficiently solve computational search problems [Handbook of Metaheuristics]. There might be multiple heuristics for solving a problem. Heuristics have their own strength and weakness. In this project, we want to use machine-learning techniques to learn the strength and weakness of each heuristic while we are using them in an iterative search for finding high quality solutions and then use them intelligently for the rest of the search. Once a new information is gathered during the search the hyper-heuristic algorithm automatically adjusts the heuristics.

Advisor: Ahmad Hemmati

Machine learning for solving satisfiability problems and applications in cryptanalysis

Advisor: Igor Semaev

Hybrid modeling approaches for well drilling with Sintef

Several topics are available.

"Flow models" are first-principles models simulating the flow, temperature and pressure in a well being drilled. Our project is exploring "hybrid approaches" where these models are combined with machine learning models that either learn from time series data from flow model runs or from real-world measurements during drilling. The goal is to better detect drilling problems such as hole cleaning, make more accurate predictions and correctly learn from and interpret real-word data.

The "surrogate model" refers to  a ML model which learns to mimic the flow model by learning from the model inputs and outputs. Use cases for surrogate models include model predictions where speed is favoured over accuracy and exploration of parameter space.

Surrogate models with active Learning

While it is possible to produce a nearly unlimited amount of training data by running the flow model, the surrogate model may still perform poorly if it lacks training data in the part of the parameter space it operates in or if it "forgets" areas of the parameter space by being fed too much data from a narrow range of parameters.

The goal of this thesis is to build a surrogate model (with any architecture) for some restricted parameter range and implement an active learning approach where the ML requests more model runs from the flow model in the parts of the parameter space where it is needed the most. The end result should be a surrogate model that is quick and performs acceptably well over the whole defined parameter range.

Surrogate models trained via adversarial learning

How best to train surrogate models from runs of the flow model is an open question. This master thesis would use the adversarial learning approach to build a surrogate model which to its "adversary" becomes indistinguishable from the output of an actual flow model run.

GPU-based Surrogate models for parameter search

While CPU speed largely stalled 20 years ago in terms of working frequency on single cores, multi-core CPUs and especially GPUs took off and delivered increases in computational power by parallelizing computations.

Modern machine learning such as deep learning takes advantage this boom in computing power by running on GPUs.

The SINTEF flow models in contrast, are software programs that runs on a CPU and does not happen to utilize multi-core CPU functionality. The model runs advance time-step by time-step and each time step relies on the results from the previous time step. The flow models are therefore fundamentally sequential and not well suited to massive parallelization.

It is however of interest to run different model runs in parallel, to explore parameter spaces. The use cases for this includes model calibration, problem detection and hypothesis generation and testing.

The task of this thesis is to implement an ML-based surrogate model in such a way that many surrogate model outputs can be produced at the same time using a single GPU. This will likely entail some trade off with model size and maybe some coding tricks.

Uncertainty estimates of hybrid predictions (Lots of room for creativity, might need to steer it more, needs good background literature)

When using predictions from a ML model trained on time series data, it is useful to know if it's accurate or should be trusted. The student is challenged to develop hybrid approaches that incorporates estimates of uncertainty. Components could include reporting variance from ML ensembles trained on a diversity of time series data, implementation of conformal predictions, analysis of training data parameter ranges vs current input, etc. The output should be a "traffic light signal" roughly indicating the accuracy of the predictions.

Transfer learning approaches

We're assuming an ML model is to be used for time series prediction

It is possible to train an ML on a wide range of scenarios in the flow models, but we expect that to perform well, the model also needs to see model runs representative of the type of well and drilling operation it will be used in. In this thesis the student implements a transfer learning approach, where the model is trained on general model runs and fine-tuned on a most representative data set.

(Bonus1: implementing one-shot learning, Bonus2: Using real-world data in the fine-tuning stage)

ML capable of reframing situations

When a human oversees an operation like well drilling, she has a mental model of the situation and new data such as pressure readings from the well is interpreted in light of this model. This is referred to as "framing" and is the normal mode of work. However, when a problem occurs, it becomes harder to reconcile the data with the mental model. The human then goes into "reframing", building a new mental model that includes the ongoing problem. This can be seen as a process of hypothesis generation and testing.

A computer model however, lacks re-framing. A flow model will keep making predictions under the assumption of no problems and a separate alarm system will use the deviation between the model predictions and reality to raise an alarm. This is in a sense how all alarm systems work, but it means that the human must discard the computer model as a tool at the same time as she's handling a crisis.

The student is given access to a flow model and a surrogate model which can learn from model runs both with and without hole cleaning and is challenged to develop a hybrid approach where the ML+flow model continuously performs hypothesis generation and testing and is able to "switch" into predictions of  a hole cleaning problem and different remediations of this.

Advisor: Philippe Nivlet at Sintef together with advisor from UiB

Explainable AI at Equinor

In the project Machine Teaching for XAI (see  https://xai.w.uib.no ) a master thesis in collaboration between UiB and Equinor.

Advisor: One of Pekka Parviainen/Jan Arne Telle/Emmanuel Arrighi + Bjarte Johansen from Equinor.

Explainable AI at Eviny

In the project Machine Teaching for XAI (see  https://xai.w.uib.no ) a master thesis in collaboration between UiB and Eviny.

Advisor: One of Pekka Parviainen/Jan Arne Telle/Emmanuel Arrighi + Kristian Flikka from Eviny.

If you want to suggest your own topic, please contact Pekka Parviainen ,  Fabio Massimo Zennaro or Nello Blaser .

Graph

youtube logo

The Future of AI Research: 20 Thesis Ideas for Undergraduate Students in Machine Learning and Deep Learning for 2023!

A comprehensive guide for crafting an original and innovative thesis in the field of ai..

By Aarafat Islam

“The beauty of machine learning is that it can be applied to any problem you want to solve, as long as you can provide the computer with enough examples.” — Andrew Ng

This article provides a list of 20 potential thesis ideas for an undergraduate program in machine learning and deep learning in 2023. Each thesis idea includes an  introduction , which presents a brief overview of the topic and the  research objectives . The ideas provided are related to different areas of machine learning and deep learning, such as computer vision, natural language processing, robotics, finance, drug discovery, and more. The article also includes explanations, examples, and conclusions for each thesis idea, which can help guide the research and provide a clear understanding of the potential contributions and outcomes of the proposed research. The article also emphasized the importance of originality and the need for proper citation in order to avoid plagiarism.

1. Investigating the use of Generative Adversarial Networks (GANs) in medical imaging:  A deep learning approach to improve the accuracy of medical diagnoses.

Introduction:  Medical imaging is an important tool in the diagnosis and treatment of various medical conditions. However, accurately interpreting medical images can be challenging, especially for less experienced doctors. This thesis aims to explore the use of GANs in medical imaging, in order to improve the accuracy of medical diagnoses.

2. Exploring the use of deep learning in natural language generation (NLG): An analysis of the current state-of-the-art and future potential.

Introduction:  Natural language generation is an important field in natural language processing (NLP) that deals with creating human-like text automatically. Deep learning has shown promising results in NLP tasks such as machine translation, sentiment analysis, and question-answering. This thesis aims to explore the use of deep learning in NLG and analyze the current state-of-the-art models, as well as potential future developments.

3. Development and evaluation of deep reinforcement learning (RL) for robotic navigation and control.

Introduction:  Robotic navigation and control are challenging tasks, which require a high degree of intelligence and adaptability. Deep RL has shown promising results in various robotics tasks, such as robotic arm control, autonomous navigation, and manipulation. This thesis aims to develop and evaluate a deep RL-based approach for robotic navigation and control and evaluate its performance in various environments and tasks.

4. Investigating the use of deep learning for drug discovery and development.

Introduction:  Drug discovery and development is a time-consuming and expensive process, which often involves high failure rates. Deep learning has been used to improve various tasks in bioinformatics and biotechnology, such as protein structure prediction and gene expression analysis. This thesis aims to investigate the use of deep learning for drug discovery and development and examine its potential to improve the efficiency and accuracy of the drug development process.

5. Comparison of deep learning and traditional machine learning methods for anomaly detection in time series data.

Introduction:  Anomaly detection in time series data is a challenging task, which is important in various fields such as finance, healthcare, and manufacturing. Deep learning methods have been used to improve anomaly detection in time series data, while traditional machine learning methods have been widely used as well. This thesis aims to compare deep learning and traditional machine learning methods for anomaly detection in time series data and examine their respective strengths and weaknesses.

machine learning thesis topics

Photo by  Joanna Kosinska  on  Unsplash

6. Use of deep transfer learning in speech recognition and synthesis.

Introduction:  Speech recognition and synthesis are areas of natural language processing that focus on converting spoken language to text and vice versa. Transfer learning has been widely used in deep learning-based speech recognition and synthesis systems to improve their performance by reusing the features learned from other tasks. This thesis aims to investigate the use of transfer learning in speech recognition and synthesis and how it improves the performance of the system in comparison to traditional methods.

7. The use of deep learning for financial prediction.

Introduction:  Financial prediction is a challenging task that requires a high degree of intelligence and adaptability, especially in the field of stock market prediction. Deep learning has shown promising results in various financial prediction tasks, such as stock price prediction and credit risk analysis. This thesis aims to investigate the use of deep learning for financial prediction and examine its potential to improve the accuracy of financial forecasting.

8. Investigating the use of deep learning for computer vision in agriculture.

Introduction:  Computer vision has the potential to revolutionize the field of agriculture by improving crop monitoring, precision farming, and yield prediction. Deep learning has been used to improve various computer vision tasks, such as object detection, semantic segmentation, and image classification. This thesis aims to investigate the use of deep learning for computer vision in agriculture and examine its potential to improve the efficiency and accuracy of crop monitoring and precision farming.

9. Development and evaluation of deep learning models for generative design in engineering and architecture.

Introduction:  Generative design is a powerful tool in engineering and architecture that can help optimize designs and reduce human error. Deep learning has been used to improve various generative design tasks, such as design optimization and form generation. This thesis aims to develop and evaluate deep learning models for generative design in engineering and architecture and examine their potential to improve the efficiency and accuracy of the design process.

10. Investigating the use of deep learning for natural language understanding.

Introduction:  Natural language understanding is a complex task of natural language processing that involves extracting meaning from text. Deep learning has been used to improve various NLP tasks, such as machine translation, sentiment analysis, and question-answering. This thesis aims to investigate the use of deep learning for natural language understanding and examine its potential to improve the efficiency and accuracy of natural language understanding systems.

machine learning thesis topics

Photo by  UX Indonesia  on  Unsplash

11. Comparing deep learning and traditional machine learning methods for image compression.

Introduction:  Image compression is an important task in image processing and computer vision. It enables faster data transmission and storage of image files. Deep learning methods have been used to improve image compression, while traditional machine learning methods have been widely used as well. This thesis aims to compare deep learning and traditional machine learning methods for image compression and examine their respective strengths and weaknesses.

12. Using deep learning for sentiment analysis in social media.

Introduction:  Sentiment analysis in social media is an important task that can help businesses and organizations understand their customers’ opinions and feedback. Deep learning has been used to improve sentiment analysis in social media, by training models on large datasets of social media text. This thesis aims to use deep learning for sentiment analysis in social media, and evaluate its performance against traditional machine learning methods.

13. Investigating the use of deep learning for image generation.

Introduction:  Image generation is a task in computer vision that involves creating new images from scratch or modifying existing images. Deep learning has been used to improve various image generation tasks, such as super-resolution, style transfer, and face generation. This thesis aims to investigate the use of deep learning for image generation and examine its potential to improve the quality and diversity of generated images.

14. Development and evaluation of deep learning models for anomaly detection in cybersecurity.

Introduction:  Anomaly detection in cybersecurity is an important task that can help detect and prevent cyber-attacks. Deep learning has been used to improve various anomaly detection tasks, such as intrusion detection and malware detection. This thesis aims to develop and evaluate deep learning models for anomaly detection in cybersecurity and examine their potential to improve the efficiency and accuracy of cybersecurity systems.

15. Investigating the use of deep learning for natural language summarization.

Introduction:  Natural language summarization is an important task in natural language processing that involves creating a condensed version of a text that preserves its main meaning. Deep learning has been used to improve various natural language summarization tasks, such as document summarization and headline generation. This thesis aims to investigate the use of deep learning for natural language summarization and examine its potential to improve the efficiency and accuracy of natural language summarization systems.

machine learning thesis topics

Photo by  Windows  on  Unsplash

16. Development and evaluation of deep learning models for facial expression recognition.

Introduction:  Facial expression recognition is an important task in computer vision and has many practical applications, such as human-computer interaction, emotion recognition, and psychological studies. Deep learning has been used to improve facial expression recognition, by training models on large datasets of images. This thesis aims to develop and evaluate deep learning models for facial expression recognition and examine their performance against traditional machine learning methods.

17. Investigating the use of deep learning for generative models in music and audio.

Introduction:  Music and audio synthesis is an important task in audio processing, which has many practical applications, such as music generation and speech synthesis. Deep learning has been used to improve generative models for music and audio, by training models on large datasets of audio data. This thesis aims to investigate the use of deep learning for generative models in music and audio and examine its potential to improve the quality and diversity of generated audio.

18. Study the comparison of deep learning models with traditional algorithms for anomaly detection in network traffic.

Introduction:  Anomaly detection in network traffic is an important task that can help detect and prevent cyber-attacks. Deep learning models have been used for this task, and traditional methods such as clustering and rule-based systems are widely used as well. This thesis aims to compare deep learning models with traditional algorithms for anomaly detection in network traffic and analyze the trade-offs between the models in terms of accuracy and scalability.

19. Investigating the use of deep learning for improving recommender systems.

Introduction:  Recommender systems are widely used in many applications such as online shopping, music streaming, and movie streaming. Deep learning has been used to improve the performance of recommender systems, by training models on large datasets of user-item interactions. This thesis aims to investigate the use of deep learning for improving recommender systems and compare its performance with traditional content-based and collaborative filtering approaches.

20. Development and evaluation of deep learning models for multi-modal data analysis.

Introduction:  Multi-modal data analysis is the task of analyzing and understanding data from multiple sources such as text, images, and audio. Deep learning has been used to improve multi-modal data analysis, by training models on large datasets of multi-modal data. This thesis aims to develop and evaluate deep learning models for multi-modal data analysis and analyze their potential to improve performance in comparison to single-modal models.

I hope that this article has provided you with a useful guide for your thesis research in machine learning and deep learning. Remember to conduct a thorough literature review and to include proper citations in your work, as well as to be original in your research to avoid plagiarism. I wish you all the best of luck with your thesis and your research endeavors!

Continue Learning

Using chains and agents for llm application development.

Step-by-step guide to using chains and agents in LangChain

Your Local LLM using FastAPI

FastAPI is a modern, fast, and easy-to-use web framework for building APIs with Python. It is based on the standard Python pointer type and supports features such as data validation, documentation…

Pooling Layer — Short and Simple

Here's all the information you should know about Pooling Layer in CNN

2050: What AI Foresees for the Future World

Exploring the future of AI and its potential impact on various aspects of our world by the year 2050.

Role of Artificial Intelligence in Metaverse

Exploring the Saga of Metaverse with AI

Mastering the GPT-3 Temperature Parameter with Ruby

CodeAvail

Exploring 250+ Machine Learning Research Topics

machine learning research topics

In recent years, machine learning has become super popular and grown very quickly. This happened because technology got better, and there’s a lot more data available. Because of this, we’ve seen lots of new and amazing things happen in different areas. Machine learning research is what makes all these cool things possible. In this blog, we’ll talk about machine learning research topics, why they’re important, how you can pick one, what areas are popular to study, what’s new and exciting, the tough problems, and where you can find help if you want to be a researcher.

Why Does Machine Learning Research Matter?

Table of Contents

Machine learning research is at the heart of the AI revolution. It underpins the development of intelligent systems capable of making predictions, automating tasks, and improving decision-making across industries. The importance of this research can be summarized as follows:

Advancements in Technology

The growth of machine learning research has led to the development of powerful algorithms, tools, and frameworks. Numerous industries, including healthcare, banking, autonomous cars, and natural language processing, have found use for these technology.

As researchers continue to push the boundaries of what’s possible, we can expect even more transformative technologies to emerge.

Real-world Applications

Machine learning research has brought about tangible changes in our daily lives. Voice assistants like Siri and Alexa, recommendation systems on streaming platforms, and personalized healthcare diagnostics are just a few examples of how this research impacts our world. 

By working on new research topics, scientists can further refine these applications and create new ones.

Economic and Industrial Impacts

The economic implications of machine learning research are substantial. Companies that harness the power of machine learning gain a competitive edge in the market. 

This creates a demand for skilled machine learning researchers, driving job opportunities and contributing to economic growth.

How to Choose the Machine Learning Research Topics?

Selecting the right machine learning research topics is crucial for your success as a machine learning researcher. Here’s a guide to help you make an informed decision:

  • Understanding Your Interests

Start by considering your personal interests. Machine learning is a broad field with applications in virtually every sector. By choosing a topic that aligns with your passions, you’ll stay motivated and engaged throughout your research journey.

  • Reviewing Current Trends

Stay updated on the latest trends in machine learning. Attend conferences, read research papers, and engage with the community to identify emerging research topics. Current trends often lead to exciting breakthroughs.

  • Identifying Gaps in Existing Research

Sometimes, the most promising research topics involve addressing gaps in existing knowledge. These gaps may become evident through your own experiences, discussions with peers, or in the course of your studies.

  • Collaborating with Experts

Collaboration is key in research. Working with experts in the field can help you refine your research topic and gain valuable insights. Seek mentors and collaborators who can guide you.

250+ Machine Learning Research Topics: Category-wise

Supervised learning.

  • Explainable AI for Decision Support
  • Few-shot Learning Methods
  • Time Series Forecasting with Deep Learning
  • Handling Imbalanced Datasets in Classification
  • Regression Techniques for Non-linear Data
  • Transfer Learning in Supervised Settings
  • Multi-label Classification Strategies
  • Semi-Supervised Learning Approaches
  • Novel Feature Selection Methods
  • Anomaly Detection in Supervised Scenarios
  • Federated Learning for Distributed Supervised Models
  • Ensemble Learning for Improved Accuracy
  • Automated Hyperparameter Tuning
  • Ethical Implications in Supervised Models
  • Interpretability of Deep Neural Networks.

Unsupervised Learning

  • Unsupervised Clustering of High-dimensional Data
  • Semi-Supervised Clustering Approaches
  • Density Estimation in Unsupervised Learning
  • Anomaly Detection in Unsupervised Settings
  • Transfer Learning for Unsupervised Tasks
  • Representation Learning in Unsupervised Learning
  • Outlier Detection Techniques
  • Generative Models for Data Synthesis
  • Manifold Learning in High-dimensional Spaces
  • Unsupervised Feature Selection
  • Privacy-Preserving Unsupervised Learning
  • Community Detection in Complex Networks
  • Clustering Interpretability and Visualization
  • Unsupervised Learning for Image Segmentation
  • Autoencoders for Dimensionality Reduction.

Reinforcement Learning

  • Deep Reinforcement Learning in Real-world Applications
  • Safe Reinforcement Learning for Autonomous Systems
  • Transfer Learning in Reinforcement Learning
  • Imitation Learning and Apprenticeship Learning
  • Multi-agent Reinforcement Learning
  • Explainable Reinforcement Learning Policies
  • Hierarchical Reinforcement Learning
  • Model-based Reinforcement Learning
  • Curriculum Learning in Reinforcement Learning
  • Reinforcement Learning in Robotics
  • Exploration vs. Exploitation Strategies
  • Reward Function Design and Ethical Considerations
  • Reinforcement Learning in Healthcare
  • Continuous Action Spaces in RL
  • Reinforcement Learning for Resource Management.

Natural Language Processing (NLP)

  • Multilingual and Cross-lingual NLP
  • Contextualized Word Embeddings
  • Bias Detection and Mitigation in NLP
  • Named Entity Recognition for Low-resource Languages
  • Sentiment Analysis in Social Media Text
  • Dialogue Systems for Improved Customer Service
  • Text Summarization for News Articles
  • Low-resource Machine Translation
  • Explainable NLP Models
  • Coreference Resolution in NLP
  • Question Answering in Specific Domains
  • Detecting Fake News and Misinformation
  • NLP for Healthcare: Clinical Document Understanding
  • Emotion Analysis in Text
  • Text Generation with Controlled Attributes.

Computer Vision

  • Video Action Recognition and Event Detection
  • Object Detection in Challenging Conditions (e.g., low light)
  • Explainable Computer Vision Models
  • Image Captioning for Accessibility
  • Large-scale Image Retrieval
  • Domain Adaptation in Computer Vision
  • Fine-grained Image Classification
  • Facial Expression Recognition
  • Visual Question Answering
  • Self-supervised Learning for Visual Representations
  • Weakly Supervised Object Localization
  • Human Pose Estimation in 3D
  • Scene Understanding in Autonomous Vehicles
  • Image Super-resolution
  • Gaze Estimation for Human-Computer Interaction.

Deep Learning

  • Neural Architecture Search for Efficient Models
  • Self-attention Mechanisms and Transformers
  • Interpretability in Deep Learning Models
  • Robustness of Deep Neural Networks
  • Generative Adversarial Networks (GANs) for Data Augmentation
  • Neural Style Transfer in Art and Design
  • Adversarial Attacks and Defenses
  • Neural Networks for Audio and Speech Processing
  • Explainable AI for Healthcare Diagnosis
  • Automated Machine Learning (AutoML)
  • Reinforcement Learning with Deep Neural Networks
  • Model Compression and Quantization
  • Lifelong Learning with Deep Learning Models
  • Multimodal Learning with Vision and Language
  • Federated Learning for Privacy-preserving Deep Learning.

Explainable AI

  • Visualizing Model Decision Boundaries
  • Saliency Maps and Feature Attribution
  • Rule-based Explanations for Black-box Models
  • Contrastive Explanations for Model Interpretability
  • Counterfactual Explanations and What-if Analysis
  • Human-centered AI for Explainable Healthcare
  • Ethics and Fairness in Explainable AI
  • Explanation Generation for Natural Language Processing
  • Explainable AI in Financial Risk Assessment
  • User-friendly Interfaces for Model Interpretability
  • Scalability and Efficiency in Explainable Models
  • Hybrid Models for Combined Accuracy and Explainability
  • Post-hoc vs. Intrinsic Explanations
  • Evaluation Metrics for Explanation Quality
  • Explainable AI for Autonomous Vehicles.

Transfer Learning

  • Zero-shot Learning and Few-shot Learning
  • Cross-domain Transfer Learning
  • Domain Adaptation for Improved Generalization
  • Multilingual Transfer Learning in NLP
  • Pretraining and Fine-tuning Techniques
  • Lifelong Learning and Continual Learning
  • Domain-specific Transfer Learning Applications
  • Model Distillation for Knowledge Transfer
  • Contrastive Learning for Transfer Learning
  • Self-training and Pseudo-labeling
  • Dynamic Adaption of Pretrained Models
  • Privacy-Preserving Transfer Learning
  • Unsupervised Domain Adaptation
  • Negative Transfer Avoidance in Transfer Learning.

Federated Learning

  • Secure Aggregation in Federated Learning
  • Communication-efficient Federated Learning
  • Privacy-preserving Techniques in Federated Learning
  • Federated Transfer Learning
  • Heterogeneous Federated Learning
  • Real-world Applications of Federated Learning
  • Federated Learning for Edge Devices
  • Federated Learning for Healthcare Data
  • Differential Privacy in Federated Learning
  • Byzantine-robust Federated Learning
  • Federated Learning with Non-IID Data
  • Model Selection in Federated Learning
  • Scalable Federated Learning for Large Datasets
  • Client Selection and Sampling Strategies
  • Global Model Update Synchronization in Federated Learning.

Quantum Machine Learning

  • Quantum Neural Networks and Quantum Circuit Learning
  • Quantum-enhanced Optimization for Machine Learning
  • Quantum Data Compression and Quantum Principal Component Analysis
  • Quantum Kernels and Quantum Feature Maps
  • Quantum Variational Autoencoders
  • Quantum Transfer Learning
  • Quantum-inspired Classical Algorithms for ML
  • Hybrid Quantum-Classical Models
  • Quantum Machine Learning on Near-term Quantum Devices
  • Quantum-inspired Reinforcement Learning
  • Quantum Computing for Quantum Chemistry and Drug Discovery
  • Quantum Machine Learning for Finance
  • Quantum Data Structures and Quantum Databases
  • Quantum-enhanced Cryptography in Machine Learning
  • Quantum Generative Models and Quantum GANs.

Ethical AI and Bias Mitigation

  • Fairness-aware Machine Learning Algorithms
  • Bias Detection and Mitigation in Real-world Data
  • Explainable AI for Ethical Decision Support
  • Algorithmic Accountability and Transparency
  • Privacy-preserving AI and Data Governance
  • Ethical Considerations in AI for Healthcare
  • Fairness in Recommender Systems
  • Bias and Fairness in NLP Models
  • Auditing AI Systems for Bias
  • Societal Implications of AI in Criminal Justice
  • Ethical AI Education and Training
  • Bias Mitigation in Autonomous Vehicles
  • Fair AI in Financial and Hiring Decisions
  • Case Studies in Ethical AI Failures
  • Legal and Policy Frameworks for Ethical AI.

Meta-Learning and AutoML

  • Neural Architecture Search (NAS) for Efficient Models
  • Transfer Learning in NAS
  • Reinforcement Learning for NAS
  • Multi-objective NAS
  • Automated Data Augmentation
  • Neural Architecture Optimization for Edge Devices
  • Bayesian Optimization for AutoML
  • Model Compression and Quantization in AutoML
  • AutoML for Federated Learning
  • AutoML in Healthcare Diagnostics
  • Explainable AutoML
  • Cost-sensitive Learning in AutoML
  • AutoML for Small Data
  • Human-in-the-Loop AutoML.

AI for Healthcare and Medicine

  • Disease Prediction and Early Diagnosis
  • Medical Image Analysis with Deep Learning
  • Drug Discovery and Molecular Modeling
  • Electronic Health Record Analysis
  • Predictive Analytics in Healthcare
  • Personalized Treatment Planning
  • Healthcare Fraud Detection
  • Telemedicine and Remote Patient Monitoring
  • AI in Radiology and Pathology
  • AI in Drug Repurposing
  • AI for Medical Robotics and Surgery
  • Genomic Data Analysis
  • AI-powered Mental Health Assessment
  • Explainable AI in Healthcare Decision Support
  • AI in Epidemiology and Outbreak Prediction.

AI in Finance and Investment

  • Algorithmic Trading and High-frequency Trading
  • Credit Scoring and Risk Assessment
  • Fraud Detection and Anti-money Laundering
  • Portfolio Optimization with AI
  • Financial Market Prediction
  • Sentiment Analysis in Financial News
  • Explainable AI in Financial Decision-making
  • Algorithmic Pricing and Dynamic Pricing Strategies
  • AI in Cryptocurrency and Blockchain
  • Customer Behavior Analysis in Banking
  • Explainable AI in Credit Decisioning
  • AI in Regulatory Compliance
  • Ethical AI in Financial Services
  • AI for Real Estate Investment
  • Automated Financial Reporting.

AI in Climate Change and Sustainability

  • Climate Modeling and Prediction
  • Renewable Energy Forecasting
  • Smart Grid Optimization
  • Energy Consumption Forecasting
  • Carbon Emission Reduction with AI
  • Ecosystem Monitoring and Preservation
  • Precision Agriculture with AI
  • AI for Wildlife Conservation
  • Natural Disaster Prediction and Management
  • Water Resource Management with AI
  • Sustainable Transportation and Urban Planning
  • Climate Change Mitigation Strategies with AI
  • Environmental Impact Assessment with Machine Learning
  • Eco-friendly Supply Chain Optimization
  • Ethical AI in Climate-related Decision Support.

Data Privacy and Security

  • Differential Privacy Mechanisms
  • Federated Learning for Privacy-preserving AI
  • Secure Multi-Party Computation
  • Privacy-enhancing Technologies in Machine Learning
  • Homomorphic Encryption for Machine Learning
  • Ethical Considerations in Data Privacy
  • Privacy-preserving AI in Healthcare
  • AI for Secure Authentication and Access Control
  • Blockchain and AI for Data Security
  • Explainable Privacy in Machine Learning
  • Privacy-preserving AI in Government and Public Services
  • Privacy-compliant AI for IoT and Edge Devices
  • Secure AI Models Sharing and Deployment
  • Privacy-preserving AI in Financial Transactions
  • AI in the Legal Frameworks of Data Privacy.

Global Collaboration in Research

  • International Research Partnerships and Collaboration Models
  • Multilingual and Cross-cultural AI Research
  • Addressing Global Healthcare Challenges with AI
  • Ethical Considerations in International AI Collaborations
  • Interdisciplinary AI Research in Global Challenges
  • AI Ethics and Human Rights in Global Research
  • Data Sharing and Data Access in Global AI Research
  • Cross-border Research Regulations and Compliance
  • AI Innovation Hubs and International Research Centers
  • AI Education and Training for Global Communities
  • Humanitarian AI and AI for Sustainable Development Goals
  • AI for Cultural Preservation and Heritage Protection
  • Collaboration in AI-related Global Crises
  • AI in Cross-cultural Communication and Understanding
  • Global AI for Environmental Sustainability and Conservation.

Emerging Trends and Hot Topics in Machine Learning Research

The landscape of machine learning research topics is constantly evolving. Here are some of the emerging trends and hot topics that are shaping the field:

As AI systems become more prevalent, addressing ethical concerns and mitigating bias in algorithms are critical research areas.

Interpretable and Explainable Models

Understanding why machine learning models make specific decisions is crucial for their adoption in sensitive areas, such as healthcare and finance.

Meta-learning algorithms are designed to enable machines to learn how to learn, while AutoML aims to automate the machine learning process itself.

Machine learning is revolutionizing the healthcare sector, from diagnostic tools to drug discovery and patient care.

Algorithmic trading, risk assessment, and fraud detection are just a few applications of AI in finance, creating a wealth of research opportunities.

Machine learning research is crucial in analyzing and mitigating the impacts of climate change and promoting sustainable practices.

Challenges and Future Directions

While machine learning research has made tremendous strides, it also faces several challenges:

  • Data Privacy and Security: As machine learning models require vast amounts of data, protecting individual privacy and data security are paramount concerns.
  • Scalability and Efficiency: Developing efficient algorithms that can handle increasingly large datasets and complex computations remains a challenge.
  • Ensuring Fairness and Transparency: Addressing bias in machine learning models and making their decisions transparent is essential for equitable AI systems.
  • Quantum Computing and Machine Learning: The integration of quantum computing and machine learning has the potential to revolutionize the field, but it also presents unique challenges.
  • Global Collaboration in Research: Machine learning research benefits from collaboration on a global scale. Ensuring that researchers from diverse backgrounds work together is vital for progress.

Resources for Machine Learning Researchers

If you’re looking to embark on a journey in machine learning research topics, there are various resources at your disposal:

  • Journals and Conferences

Journals such as the “Journal of Machine Learning Research” and conferences like NeurIPS and ICML provide a platform for publishing and discussing research findings.

  • Online Communities and Forums

Platforms like Stack Overflow, GitHub, and dedicated forums for machine learning provide spaces for collaboration and problem-solving.

  • Datasets and Tools

Open-source datasets and tools like TensorFlow and PyTorch simplify the research process by providing access to data and pre-built models.

  • Research Grants and Funding Opportunities

Many organizations and government agencies offer research grants and funding for machine learning projects. Seek out these opportunities to support your research.

Machine learning research is like a superhero in the world of technology. To be a part of this exciting journey, it’s important to choose the right machine learning research topics and keep up with the latest trends.

Machine learning research makes our lives better. It powers things like smart assistants and life-saving medical tools. It’s like the force driving the future of technology and society.

But, there are challenges too. We need to work together and be ethical in our research. Everyone should benefit from this technology. The future of machine learning research is incredibly bright. If you want to be a part of it, get ready for an exciting adventure. You can help create new solutions and make a big impact on the world.

Related Posts

Tips on How To Tackle A Machine Learning Project As A Beginner

Tips on How To Tackle A Machine Learning Project As A Beginner

Here in this blog, CodeAvail experts will explain to you tips on how to tackle a machine learning project as a beginner step by step…

Artificial Intelligence and Machine Learning Basics for Beginners

Artificial Intelligence and Machine Learning Basics for Beginners

Here in this blog, CodeAvail experts will explain to you Artificial Intelligence and Machine Learning basics for beginners in detail step by step. What is…

Machine Learning - CMU

PhD Dissertations

PhD Dissertations

[all are .pdf files].

Reliable and Practical Machine Learning for Dynamic Healthcare Settings Helen Zhou, 2023

Automatic customization of large-scale spiking network models to neuronal population activity (unavailable) Shenghao Wu, 2023

Estimation of BVk functions from scattered data (unavailable) Addison J. Hu, 2023

Rethinking object categorization in computer vision (unavailable) Jayanth Koushik, 2023

Advances in Statistical Gene Networks Jinjin Tian, 2023 Post-hoc calibration without distributional assumptions Chirag Gupta, 2023

The Role of Noise, Proxies, and Dynamics in Algorithmic Fairness Nil-Jana Akpinar, 2023

Collaborative learning by leveraging siloed data Sebastian Caldas, 2023

Modeling Epidemiological Time Series Aaron Rumack, 2023

Human-Centered Machine Learning: A Statistical and Algorithmic Perspective Leqi Liu, 2023

Uncertainty Quantification under Distribution Shifts Aleksandr Podkopaev, 2023

Probabilistic Reinforcement Learning: Using Data to Define Desired Outcomes, and Inferring How to Get There Benjamin Eysenbach, 2023

Comparing Forecasters and Abstaining Classifiers Yo Joong Choe, 2023

Using Task Driven Methods to Uncover Representations of Human Vision and Semantics Aria Yuan Wang, 2023

Data-driven Decisions - An Anomaly Detection Perspective Shubhranshu Shekhar, 2023

Applied Mathematics of the Future Kin G. Olivares, 2023

METHODS AND APPLICATIONS OF EXPLAINABLE MACHINE LEARNING Joon Sik Kim, 2023

NEURAL REASONING FOR QUESTION ANSWERING Haitian Sun, 2023

Principled Machine Learning for Societally Consequential Decision Making Amanda Coston, 2023

Long term brain dynamics extend cognitive neuroscience to timescales relevant for health and physiology Maxwell B. Wang

Long term brain dynamics extend cognitive neuroscience to timescales relevant for health and physiology Darby M. Losey, 2023

Calibrated Conditional Density Models and Predictive Inference via Local Diagnostics David Zhao, 2023

Towards an Application-based Pipeline for Explainability Gregory Plumb, 2022

Objective Criteria for Explainable Machine Learning Chih-Kuan Yeh, 2022

Making Scientific Peer Review Scientific Ivan Stelmakh, 2022

Facets of regularization in high-dimensional learning: Cross-validation, risk monotonization, and model complexity Pratik Patil, 2022

Active Robot Perception using Programmable Light Curtains Siddharth Ancha, 2022

Strategies for Black-Box and Multi-Objective Optimization Biswajit Paria, 2022

Unifying State and Policy-Level Explanations for Reinforcement Learning Nicholay Topin, 2022

Sensor Fusion Frameworks for Nowcasting Maria Jahja, 2022

Equilibrium Approaches to Modern Deep Learning Shaojie Bai, 2022

Towards General Natural Language Understanding with Probabilistic Worldbuilding Abulhair Saparov, 2022

Applications of Point Process Modeling to Spiking Neurons (Unavailable) Yu Chen, 2021

Neural variability: structure, sources, control, and data augmentation Akash Umakantha, 2021

Structure and time course of neural population activity during learning Jay Hennig, 2021

Cross-view Learning with Limited Supervision Yao-Hung Hubert Tsai, 2021

Meta Reinforcement Learning through Memory Emilio Parisotto, 2021

Learning Embodied Agents with Scalably-Supervised Reinforcement Learning Lisa Lee, 2021

Learning to Predict and Make Decisions under Distribution Shift Yifan Wu, 2021

Statistical Game Theory Arun Sai Suggala, 2021

Towards Knowledge-capable AI: Agents that See, Speak, Act and Know Kenneth Marino, 2021

Learning and Reasoning with Fast Semidefinite Programming and Mixing Methods Po-Wei Wang, 2021

Bridging Language in Machines with Language in the Brain Mariya Toneva, 2021

Curriculum Learning Otilia Stretcu, 2021

Principles of Learning in Multitask Settings: A Probabilistic Perspective Maruan Al-Shedivat, 2021

Towards Robust and Resilient Machine Learning Adarsh Prasad, 2021

Towards Training AI Agents with All Types of Experiences: A Unified ML Formalism Zhiting Hu, 2021

Building Intelligent Autonomous Navigation Agents Devendra Chaplot, 2021

Learning to See by Moving: Self-supervising 3D Scene Representations for Perception, Control, and Visual Reasoning Hsiao-Yu Fish Tung, 2021

Statistical Astrophysics: From Extrasolar Planets to the Large-scale Structure of the Universe Collin Politsch, 2020

Causal Inference with Complex Data Structures and Non-Standard Effects Kwhangho Kim, 2020

Networks, Point Processes, and Networks of Point Processes Neil Spencer, 2020

Dissecting neural variability using population recordings, network models, and neurofeedback (Unavailable) Ryan Williamson, 2020

Predicting Health and Safety: Essays in Machine Learning for Decision Support in the Public Sector Dylan Fitzpatrick, 2020

Towards a Unified Framework for Learning and Reasoning Han Zhao, 2020

Learning DAGs with Continuous Optimization Xun Zheng, 2020

Machine Learning and Multiagent Preferences Ritesh Noothigattu, 2020

Learning and Decision Making from Diverse Forms of Information Yichong Xu, 2020

Towards Data-Efficient Machine Learning Qizhe Xie, 2020

Change modeling for understanding our world and the counterfactual one(s) William Herlands, 2020

Machine Learning in High-Stakes Settings: Risks and Opportunities Maria De-Arteaga, 2020

Data Decomposition for Constrained Visual Learning Calvin Murdock, 2020

Structured Sparse Regression Methods for Learning from High-Dimensional Genomic Data Micol Marchetti-Bowick, 2020

Towards Efficient Automated Machine Learning Liam Li, 2020

LEARNING COLLECTIONS OF FUNCTIONS Emmanouil Antonios Platanios, 2020

Provable, structured, and efficient methods for robustness of deep networks to adversarial examples Eric Wong , 2020

Reconstructing and Mining Signals: Algorithms and Applications Hyun Ah Song, 2020

Probabilistic Single Cell Lineage Tracing Chieh Lin, 2020

Graphical network modeling of phase coupling in brain activity (unavailable) Josue Orellana, 2019

Strategic Exploration in Reinforcement Learning - New Algorithms and Learning Guarantees Christoph Dann, 2019 Learning Generative Models using Transformations Chun-Liang Li, 2019

Estimating Probability Distributions and their Properties Shashank Singh, 2019

Post-Inference Methods for Scalable Probabilistic Modeling and Sequential Decision Making Willie Neiswanger, 2019

Accelerating Text-as-Data Research in Computational Social Science Dallas Card, 2019

Multi-view Relationships for Analytics and Inference Eric Lei, 2019

Information flow in networks based on nonstationary multivariate neural recordings Natalie Klein, 2019

Competitive Analysis for Machine Learning & Data Science Michael Spece, 2019

The When, Where and Why of Human Memory Retrieval Qiong Zhang, 2019

Towards Effective and Efficient Learning at Scale Adams Wei Yu, 2019

Towards Literate Artificial Intelligence Mrinmaya Sachan, 2019

Learning Gene Networks Underlying Clinical Phenotypes Under SNP Perturbations From Genome-Wide Data Calvin McCarter, 2019

Unified Models for Dynamical Systems Carlton Downey, 2019

Anytime Prediction and Learning for the Balance between Computation and Accuracy Hanzhang Hu, 2019

Statistical and Computational Properties of Some "User-Friendly" Methods for High-Dimensional Estimation Alnur Ali, 2019

Nonparametric Methods with Total Variation Type Regularization Veeranjaneyulu Sadhanala, 2019

New Advances in Sparse Learning, Deep Networks, and Adversarial Learning: Theory and Applications Hongyang Zhang, 2019

Gradient Descent for Non-convex Problems in Modern Machine Learning Simon Shaolei Du, 2019

Selective Data Acquisition in Learning and Decision Making Problems Yining Wang, 2019

Anomaly Detection in Graphs and Time Series: Algorithms and Applications Bryan Hooi, 2019

Neural dynamics and interactions in the human ventral visual pathway Yuanning Li, 2018

Tuning Hyperparameters without Grad Students: Scaling up Bandit Optimisation Kirthevasan Kandasamy, 2018

Teaching Machines to Classify from Natural Language Interactions Shashank Srivastava, 2018

Statistical Inference for Geometric Data Jisu Kim, 2018

Representation Learning @ Scale Manzil Zaheer, 2018

Diversity-promoting and Large-scale Machine Learning for Healthcare Pengtao Xie, 2018

Distribution and Histogram (DIsH) Learning Junier Oliva, 2018

Stress Detection for Keystroke Dynamics Shing-Hon Lau, 2018

Sublinear-Time Learning and Inference for High-Dimensional Models Enxu Yan, 2018

Neural population activity in the visual cortex: Statistical methods and application Benjamin Cowley, 2018

Efficient Methods for Prediction and Control in Partially Observable Environments Ahmed Hefny, 2018

Learning with Staleness Wei Dai, 2018

Statistical Approach for Functionally Validating Transcription Factor Bindings Using Population SNP and Gene Expression Data Jing Xiang, 2017

New Paradigms and Optimality Guarantees in Statistical Learning and Estimation Yu-Xiang Wang, 2017

Dynamic Question Ordering: Obtaining Useful Information While Reducing User Burden Kirstin Early, 2017

New Optimization Methods for Modern Machine Learning Sashank J. Reddi, 2017

Active Search with Complex Actions and Rewards Yifei Ma, 2017

Why Machine Learning Works George D. Montañez , 2017

Source-Space Analyses in MEG/EEG and Applications to Explore Spatio-temporal Neural Dynamics in Human Vision Ying Yang , 2017

Computational Tools for Identification and Analysis of Neuronal Population Activity Pengcheng Zhou, 2016

Expressive Collaborative Music Performance via Machine Learning Gus (Guangyu) Xia, 2016

Supervision Beyond Manual Annotations for Learning Visual Representations Carl Doersch, 2016

Exploring Weakly Labeled Data Across the Noise-Bias Spectrum Robert W. H. Fisher, 2016

Optimizing Optimization: Scalable Convex Programming with Proximal Operators Matt Wytock, 2016

Combining Neural Population Recordings: Theory and Application William Bishop, 2015

Discovering Compact and Informative Structures through Data Partitioning Madalina Fiterau-Brostean, 2015

Machine Learning in Space and Time Seth R. Flaxman, 2015

The Time and Location of Natural Reading Processes in the Brain Leila Wehbe, 2015

Shape-Constrained Estimation in High Dimensions Min Xu, 2015

Spectral Probabilistic Modeling and Applications to Natural Language Processing Ankur Parikh, 2015 Computational and Statistical Advances in Testing and Learning Aaditya Kumar Ramdas, 2015

Corpora and Cognition: The Semantic Composition of Adjectives and Nouns in the Human Brain Alona Fyshe, 2015

Learning Statistical Features of Scene Images Wooyoung Lee, 2014

Towards Scalable Analysis of Images and Videos Bin Zhao, 2014

Statistical Text Analysis for Social Science Brendan T. O'Connor, 2014

Modeling Large Social Networks in Context Qirong Ho, 2014

Semi-Cooperative Learning in Smart Grid Agents Prashant P. Reddy, 2013

On Learning from Collective Data Liang Xiong, 2013

Exploiting Non-sequence Data in Dynamic Model Learning Tzu-Kuo Huang, 2013

Mathematical Theories of Interaction with Oracles Liu Yang, 2013

Short-Sighted Probabilistic Planning Felipe W. Trevizan, 2013

Statistical Models and Algorithms for Studying Hand and Finger Kinematics and their Neural Mechanisms Lucia Castellanos, 2013

Approximation Algorithms and New Models for Clustering and Learning Pranjal Awasthi, 2013

Uncovering Structure in High-Dimensions: Networks and Multi-task Learning Problems Mladen Kolar, 2013

Learning with Sparsity: Structures, Optimization and Applications Xi Chen, 2013

GraphLab: A Distributed Abstraction for Large Scale Machine Learning Yucheng Low, 2013

Graph Structured Normal Means Inference James Sharpnack, 2013 (Joint Statistics & ML PhD)

Probabilistic Models for Collecting, Analyzing, and Modeling Expression Data Hai-Son Phuoc Le, 2013

Learning Large-Scale Conditional Random Fields Joseph K. Bradley, 2013

New Statistical Applications for Differential Privacy Rob Hall, 2013 (Joint Statistics & ML PhD)

Parallel and Distributed Systems for Probabilistic Reasoning Joseph Gonzalez, 2012

Spectral Approaches to Learning Predictive Representations Byron Boots, 2012

Attribute Learning using Joint Human and Machine Computation Edith L. M. Law, 2012

Statistical Methods for Studying Genetic Variation in Populations Suyash Shringarpure, 2012

Data Mining Meets HCI: Making Sense of Large Graphs Duen Horng (Polo) Chau, 2012

Learning with Limited Supervision by Input and Output Coding Yi Zhang, 2012

Target Sequence Clustering Benjamin Shih, 2011

Nonparametric Learning in High Dimensions Han Liu, 2010 (Joint Statistics & ML PhD)

Structural Analysis of Large Networks: Observations and Applications Mary McGlohon, 2010

Modeling Purposeful Adaptive Behavior with the Principle of Maximum Causal Entropy Brian D. Ziebart, 2010

Tractable Algorithms for Proximity Search on Large Graphs Purnamrita Sarkar, 2010

Rare Category Analysis Jingrui He, 2010

Coupled Semi-Supervised Learning Andrew Carlson, 2010

Fast Algorithms for Querying and Mining Large Graphs Hanghang Tong, 2009

Efficient Matrix Models for Relational Learning Ajit Paul Singh, 2009

Exploiting Domain and Task Regularities for Robust Named Entity Recognition Andrew O. Arnold, 2009

Theoretical Foundations of Active Learning Steve Hanneke, 2009

Generalized Learning Factors Analysis: Improving Cognitive Models with Machine Learning Hao Cen, 2009

Detecting Patterns of Anomalies Kaustav Das, 2009

Dynamics of Large Networks Jurij Leskovec, 2008

Computational Methods for Analyzing and Modeling Gene Regulation Dynamics Jason Ernst, 2008

Stacked Graphical Learning Zhenzhen Kou, 2007

Actively Learning Specific Function Properties with Applications to Statistical Inference Brent Bryan, 2007

Approximate Inference, Structure Learning and Feature Estimation in Markov Random Fields Pradeep Ravikumar, 2007

Scalable Graphical Models for Social Networks Anna Goldenberg, 2007

Measure Concentration of Strongly Mixing Processes with Applications Leonid Kontorovich, 2007

Tools for Graph Mining Deepayan Chakrabarti, 2005

Automatic Discovery of Latent Variable Models Ricardo Silva, 2005

machine learning thesis topics

  • ODSC EUROPE
  • AI+ Training
  • Speak at ODSC

machine learning thesis topics

  • Data Engineering
  • Data Visualization
  • Deep Learning
  • Generative AI
  • Machine Learning
  • NLP and LLMs
  • Business & Use Cases
  • Career Advice
  • Write for us
  • ODSC Community Slack Channel
  • Upcoming Webinars

17 Compelling Machine Learning Ph.D. Dissertations

17 Compelling Machine Learning Ph.D. Dissertations

Machine Learning Modeling Research posted by Daniel Gutierrez, ODSC August 12, 2021 Daniel Gutierrez, ODSC

Working in the field of data science, I’m always seeking ways to keep current in the field and there are a number of important resources available for this purpose: new book titles, blog articles, conference sessions, Meetups, webinars/podcasts, not to mention the gems floating around in social media. But to dig even deeper, I routinely look at what’s coming out of the world’s research labs. And one great way to keep a pulse for what the research community is working on is to monitor the flow of new machine learning Ph.D. dissertations. Admittedly, many such theses are laser-focused and narrow, but from previous experience reading these documents, you can learn an awful lot about new ways to solve difficult problems over a vast range of problem domains. 

In this article, I present a number of hand-picked machine learning dissertations that I found compelling in terms of my own areas of interest and aligned with problems that I’m working on. I hope you’ll find a number of them that match your own interests. Each dissertation may be challenging to consume but the process will result in hours of satisfying summer reading. Enjoy!

Please check out my previous data science dissertation round-up article . 

1. Fitting Convex Sets to Data: Algorithms and Applications

This machine learning dissertation concerns the geometric problem of finding a convex set that best fits a given data set. The overarching question serves as an abstraction for data-analytical tasks arising in a range of scientific and engineering applications with a focus on two specific instances: (i) a key challenge that arises in solving inverse problems is ill-posedness due to a lack of measurements. A prominent family of methods for addressing such issues is based on augmenting optimization-based approaches with a convex penalty function so as to induce a desired structure in the solution. These functions are typically chosen using prior knowledge about the data. The thesis also studies the problem of learning convex penalty functions directly from data for settings in which we lack the domain expertise to choose a penalty function. The solution relies on suitably transforming the problem of learning a penalty function into a fitting task; and (ii) the problem of fitting tractably-described convex sets given the optimal value of linear functionals evaluated in different directions.

2. Structured Tensors and the Geometry of Data

This machine learning dissertation analyzes data to build a quantitative understanding of the world. Linear algebra is the foundation of algorithms, dating back one hundred years, for extracting structure from data. Modern technologies provide an abundance of multi-dimensional data, in which multiple variables or factors can be compared simultaneously. To organize and analyze such data sets we can use a tensor , the higher-order analogue of a matrix. However, many theoretical and practical challenges arise in extending linear algebra to the setting of tensors. The first part of the thesis studies and develops the algebraic theory of tensors. The second part of the thesis presents three algorithms for tensor data. The algorithms use algebraic and geometric structure to give guarantees of optimality.

3. Statistical approaches for spatial prediction and anomaly detection

This machine learning dissertation is primarily a description of three projects. It starts with a method for spatial prediction and parameter estimation for irregularly spaced, and non-Gaussian data. It is shown that by judiciously replacing the likelihood with an empirical likelihood in the Bayesian hierarchical model, approximate posterior distributions for the mean and covariance parameters can be obtained. Due to the complex nature of the hierarchical model, standard Markov chain Monte Carlo methods cannot be applied to sample from the posterior distributions. To overcome this issue, a generalized sequential Monte Carlo algorithm is used. Finally, this method is applied to iron concentrations in California. The second project focuses on anomaly detection for functional data; specifically for functional data where the observed functions may lie over different domains. By approximating each function as a low-rank sum of spline basis functions the coefficients will be compared for each basis across each function. The idea being, if two functions are similar then their respective coefficients should not be significantly different. This project concludes with an application of the proposed method to detect anomalous behavior of users of a supercomputer at NREL. The final project is an extension of the second project to two-dimensional data. This project aims to detect location and temporal anomalies from ground motion data from a fiber-optic cable using distributed acoustic sensing (DAS). 

4. Sampling for Streaming Data

Advances in data acquisition technology pose challenges in analyzing large volumes of streaming data. Sampling is a natural yet powerful tool for analyzing such data sets due to their competent estimation accuracy and low computational cost. Unfortunately, sampling methods and their statistical properties for streaming data, especially streaming time series data, are not well studied in the literature. Meanwhile, estimating the dependence structure of multidimensional streaming time-series data in real-time is challenging. With large volumes of streaming data, the problem becomes more difficult when the multidimensional data are collected asynchronously across distributed nodes, which motivates us to sample representative data points from streams. This machine learning dissertation proposes a series of leverage score-based sampling methods for streaming time series data. The simulation studies and real data analysis are conducted to validate the proposed methods. The theoretical analysis of the asymptotic behaviors of the least-squares estimator is developed based on the subsamples.

5.  Statistical Machine Learning Methods for Complex, Heterogeneous Data

This machine learning dissertation develops statistical machine learning methodology for three distinct tasks. Each method blends classical statistical approaches with machine learning methods to provide principled solutions to problems with complex, heterogeneous data sets. The first framework proposes two methods for high-dimensional shape-constrained regression and classification. These methods reshape pre-trained prediction rules to satisfy shape constraints like monotonicity and convexity. The second method provides a nonparametric approach to the econometric analysis of discrete choice. This method provides a scalable algorithm for estimating utility functions with random forests, and combines this with random effects to properly model preference heterogeneity. The final method draws inspiration from early work in statistical machine translation to construct embeddings for variable-length objects like mathematical equations

6. Topics in Multivariate Statistics with Dependent Data

This machine learning dissertation comprises four chapters. The first is an introduction to the topics of the dissertation and the remaining chapters contain the main results. Chapter 2 gives new results for consistency of maximum likelihood estimators with a focus on multivariate mixed models. The presented theory builds on the idea of using subsets of the full data to establish consistency of estimators based on the full data. The theory is applied to two multivariate mixed models for which it was unknown whether maximum likelihood estimators are consistent. In Chapter 3 an algorithm is proposed for maximum likelihood estimation of a covariance matrix when the corresponding correlation matrix can be written as the Kronecker product of two lower-dimensional correlation matrices. The proposed method is fully likelihood-based. Some desirable properties of separable correlation in comparison to separable covariance are also discussed. Chapter 4 is concerned with Bayesian vector auto-regressions (VARs). A collapsed Gibbs sampler is proposed for Bayesian VARs with predictors and the convergence properties of the algorithm are studied. 

7.  Model Selection and Estimation for High-dimensional Data Analysis

In the era of big data, uncovering useful information and hidden patterns in the data is prevalent in different fields. However, it is challenging to effectively select input variables in data and estimate their effects. The goal of this machine learning dissertation is to develop reproducible statistical approaches that provide mechanistic explanations of the phenomenon observed in big data analysis. The research contains two parts: variable selection and model estimation. The first part investigates how to measure and interpret the usefulness of an input variable using an approach called “variable importance learning” and builds tools (methodology and software) that can be widely applied. Two variable importance measures are proposed, a parametric measure SOIL and a non-parametric measure CVIL, using the idea of a model combining and cross-validation respectively. The SOIL method is theoretically shown to have the inclusion/exclusion property: When the model weights are properly around the true model, the SOIL importance can well separate the variables in the true model from the rest. The CVIL method possesses desirable theoretical properties and enhances the interpretability of many mysterious but effective machine learning methods. The second part focuses on how to estimate the effect of a useful input variable in the case where the interaction of two input variables exists. Investigated is the minimax rate of convergence for regression estimation in high-dimensional sparse linear models with two-way interactions, and construct an adaptive estimator that achieves the minimax rate of convergence regardless of the true heredity condition and the sparsity indices.

https://odsc.com/california/#register

8.  High-Dimensional Structured Regression Using Convex Optimization

While the term “Big Data” can have multiple meanings, this dissertation considers the type of data in which the number of features can be much greater than the number of observations (also known as high-dimensional data). High-dimensional data is abundant in contemporary scientific research due to the rapid advances in new data-measurement technologies and computing power. Recent advances in statistics have witnessed great development in the field of high-dimensional data analysis. This machine learning dissertation proposes three methods that study three different components of a general framework of the high-dimensional structured regression problem. A general theme of the proposed methods is that they cast a certain structured regression as a convex optimization problem. In so doing, the theoretical properties of each method can be well studied, and efficient computation is facilitated. Each method is accompanied by a thorough theoretical analysis of its performance, and also by an R package containing its practical implementation. It is shown that the proposed methods perform favorably (both theoretically and practically) compared with pre-existing methods.

9. Asymptotics and Interpretability of Decision Trees and Decision Tree Ensembles

Decision trees and decision tree ensembles are widely used nonparametric statistical models. A decision tree is a binary tree that recursively segments the covariate space along the coordinate directions to create hyper rectangles as basic prediction units for fitting constant values within each of them. A decision tree ensemble combines multiple decision trees, either in parallel or in sequence, in order to increase model flexibility and accuracy, as well as to reduce prediction variance. Despite the fact that tree models have been extensively used in practice, results on their asymptotic behaviors are scarce. This machine learning dissertation presents analyses on tree asymptotics in the perspectives of tree terminal nodes, tree ensembles, and models incorporating tree ensembles respectively. The study introduces a few new tree-related learning frameworks which provides provable statistical guarantees and interpretations. A study on the Gini index used in the greedy tree building algorithm reveals its limiting distribution, leading to the development of a test of better splitting that helps to measure the uncertain optimality of a decision tree split. This test is combined with the concept of decision tree distillation, which implements a decision tree to mimic the behavior of a block box model, to generate stable interpretations by guaranteeing a unique distillation tree structure as long as there are sufficiently many random sample points. Also applied is mild modification and regularization to the standard tree boosting to create a new boosting framework named Boulevard. Also included is an integration of two new mechanisms: honest trees , which isolate the tree terminal values from the tree structure, and adaptive shrinkage , which scales the boosting history to create an equally weighted ensemble. This theoretical development provides the prerequisite for the practice of statistical inference with boosted trees. Lastly, the thesis investigates the feasibility of incorporating existing semi-parametric models with tree boosting. 

10. Bayesian Models for Imputing Missing Data and Editing Erroneous Responses in Surveys

This dissertation develops Bayesian methods for handling unit nonresponse, item nonresponse, and erroneous responses in large-scale surveys and censuses containing categorical data. The focus is on applications of nested household data where individuals are nested within households and certain combinations of the variables are not allowed, such as the U.S. Decennial Census, as well as surveys subject to both unit and item nonresponse, such as the Current Population Survey.

11. Localized Variable Selection with Random Forest  

Due to recent advances in computer technology, the cost of collecting and storing data has dropped drastically. This makes it feasible to collect large amounts of information for each data point. This increasing trend in feature dimensionality justifies the need for research on variable selection. Random forest (RF) has demonstrated the ability to select important variables and model complex data. However, simulations confirm that it fails in detecting less influential features in presence of variables with large impacts in some cases. This dissertation proposes two algorithms for localized variable selection: clustering-based feature selection (CBFS) and locally adjusted feature importance (LAFI). Both methods aim to find regions where the effects of weaker features can be isolated and measured. CBFS combines RF variable selection with a two-stage clustering method to detect variables where their effect can be detected only in certain regions. LAFI, on the other hand, uses a binary tree approach to split data into bins based on response variable rankings, and implements RF to find important variables in each bin. Larger LAFI is assigned to variables that get selected in more bins. Simulations and real data sets are used to evaluate these variable selection methods. 

12. Functional Principal Component Analysis and Sparse Functional Regression

The focus of this dissertation is on functional data which are sparsely and irregularly observed. Such data require special consideration, as classical functional data methods and theory were developed for densely observed data. As is the case in much of functional data analysis, the functional principal components (FPCs) play a key role in current sparse functional data methods via the Karhunen-Loéve expansion. Thus, after a review of relevant background material, this dissertation is divided roughly into two parts, the first focusing specifically on theoretical properties of FPCs, and the second on regression for sparsely observed functional data.

13. Essays In Causal Inference: Addressing Bias In Observational And Randomized Studies Through Analysis And Design

In observational studies, identifying assumptions may fail, often quietly and without notice, leading to biased causal estimates. Although less of a concern in randomized trials where treatment is assigned at random, bias may still enter the equation through other means. This dissertation has three parts, each developing new methods to address a particular pattern or source of bias in the setting being studied. The first part extends the conventional sensitivity analysis methods for observational studies to better address patterns of heterogeneous confounding in matched-pair designs. The second part develops a modified difference-in-difference design for comparative interrupted time-series studies. The method permits partial identification of causal effects when the parallel trends assumption is violated by an interaction between group and history. The method is applied to a study of the repeal of Missouri’s permit-to-purchase handgun law and its effect on firearm homicide rates. The final part presents a study design to identify vaccine efficacy in randomized control trials when there is no gold standard case definition. The approach augments a two-arm randomized trial with natural variation of a genetic trait to produce a factorial experiment. 

14. Bayesian Shrinkage: Computation, Methods, and Theory

Sparsity is a standard structural assumption that is made while modeling high-dimensional statistical parameters. This assumption essentially entails a lower-dimensional embedding of the high-dimensional parameter thus enabling sound statistical inference. Apart from this obvious statistical motivation, in many modern applications of statistics such as Genomics, Neuroscience, etc. parameters of interest are indeed of this nature. For over almost two decades, spike and slab type priors have been the Bayesian gold standard for modeling of sparsity. However, due to their computational bottlenecks, shrinkage priors have emerged as a powerful alternative. This family of priors can almost exclusively be represented as a scale mixture of Gaussian distribution and posterior Markov chain Monte Carlo (MCMC) updates of related parameters are then relatively easy to design. Although shrinkage priors were tipped as having computational scalability in high-dimensions, when the number of parameters is in thousands or more, they do come with their own computational challenges. Standard MCMC algorithms implementing shrinkage priors generally scale cubic in the dimension of the parameter making real-life application of these priors severely limited. 

The first chapter of this dissertation addresses this computational issue and proposes an alternative exact posterior sampling algorithm complexity of which that linearly in the ambient dimension. The algorithm developed in the first chapter is specifically designed for regression problems. The second chapter develops a Bayesian method based on shrinkage priors for high-dimensional multiple response regression. Chapter three chooses a specific member of the shrinkage family known as the horseshoe prior and studies its convergence rates in several high-dimensional models. 

15.  Topics in Measurement Error Analysis and High-Dimensional Binary Classification

This dissertation proposes novel methods to tackle two problems: the misspecified model with measurement error and high-dimensional binary classification, both have a crucial impact on applications in public health. The first problem exists in the epidemiology practice. Epidemiologists often categorize a continuous risk predictor since categorization is thought to be more robust and interpretable, even when the true risk model is not a categorical one. Thus, their goal is to fit the categorical model and interpret the categorical parameters. The second project considers the problem of high-dimensional classification between the two groups with unequal covariance matrices. Rather than estimating the full quadratic discriminant rule, it is proposed to perform simultaneous variable selection and linear dimension reduction on original data, with the subsequent application of quadratic discriminant analysis on the reduced space. Further, in order to support the proposed methodology, two R packages were developed, CCP and DAP, along with two vignettes as long-format illustrations for their usage.

16. Model-Based Penalized Regression

This dissertation contains three chapters that consider penalized regression from a model-based perspective, interpreting penalties as assumed prior distributions for unknown regression coefficients. The first chapter shows that treating a lasso penalty as a prior can facilitate the choice of tuning parameters when standard methods for choosing the tuning parameters are not available, and when it is necessary to choose multiple tuning parameters simultaneously. The second chapter considers a possible drawback of treating penalties as models, specifically possible misspecification. The third chapter introduces structured shrinkage priors for dependent regression coefficients which generalize popular independent shrinkage priors. These can be useful in various applied settings where many regression coefficients are not only expected to be nearly or exactly equal to zero, but also structured.

17. Topics on Least Squares Estimation

This dissertation revisits and makes progress on some old but challenging problems concerning least squares estimation, the work-horse of supervised machine learning. Two major problems are addressed: (i) least squares estimation with heavy-tailed errors, and (ii) least squares estimation in non-Donsker classes. For (i), this problem is studied both from a worst-case perspective, and a more refined envelope perspective. For (ii), two case studies are performed in the context of (a) estimation involving sets and (b) estimation of multivariate isotonic functions. Understanding these particular aspects of least squares estimation problems requires several new tools in the empirical process theory, including a sharp multiplier inequality controlling the size of the multiplier empirical process, and matching upper and lower bounds for empirical processes indexed by non-Donsker classes.

How to Learn More about Machine Learning

At our upcoming event this November 16th-18th in San Francisco,  ODSC West 2021  will feature a plethora of talks, workshops, and training sessions on machine learning and machine learning research. You can  register now for 50% off all ticket types  before the discount drops to 40% in a few weeks. Some  highlighted sessions on machine learning  include:

  • Towards More Energy-Efficient Neural Networks? Use Your Brain!: Olaf de Leeuw | Data Scientist | Dataworkz
  • Practical MLOps: Automation Journey: Evgenii Vinogradov, PhD | Head of DHW Development | YooMoney
  • Applications of Modern Survival Modeling with Python: Brian Kent, PhD | Data Scientist | Founder The Crosstab Kite
  • Using Change Detection Algorithms for Detecting Anomalous Behavior in Large Systems: Veena Mendiratta, PhD | Adjunct Faculty, Network Reliability and Analytics Researcher | Northwestern University

Sessions on MLOps:

  • Tuning Hyperparameters with Reproducible Experiments: Milecia McGregor | Senior Software Engineer | Iterative
  • MLOps… From Model to Production: Filipa Peleja, PhD | Lead Data Scientist | Levi Strauss & Co
  • Operationalization of Models Developed and Deployed in Heterogeneous Platforms: Sourav Mazumder | Data Scientist, Thought Leader, AI & ML Operationalization Leader | IBM
  • Develop and Deploy a Machine Learning Pipeline in 45 Minutes with Ploomber: Eduardo Blancas | Data Scientist | Fidelity Investments

Sessions on Deep Learning:

  • GANs: Theory and Practice, Image Synthesis With GANs Using TensorFlow: Ajay Baranwal | Center Director | Center for Deep Learning in Electronic Manufacturing, Inc
  • Machine Learning With Graphs: Going Beyond Tabular Data: Dr. Clair J. Sullivan | Data Science Advocate | Neo4j
  • Deep Dive into Reinforcement Learning with PPO using TF-Agents & TensorFlow 2.0: Oliver Zeigermann | Software Developer | embarc Software Consulting GmbH
  • Get Started with Time-Series Forecasting using the Google Cloud AI Platform: Karl Weinmeister | Developer Relations Engineering Manager | Google

machine learning thesis topics

Daniel Gutierrez, ODSC

Daniel D. Gutierrez is a practicing data scientist who’s been working with data long before the field came in vogue. As a technology journalist, he enjoys keeping a pulse on this fast-paced industry. Daniel is also an educator having taught data science, machine learning and R classes at the university level. He has authored four computer industry books on database and data science technology, including his most recent title, “Machine Learning and Data Science: An Introduction to Statistical Learning Methods with R.” Daniel holds a BS in Mathematics and Computer Science from UCLA.

DE Summit Square

ODSC’s AI Weekly Recap: Week of February 23rd

AI and Data Science News posted by ODSC Team Feb 23, 2024 Open Data Science Blog Recap A new open-source Python package looks to push for accuracy and...

White House Enters Debate on “Open” Versus “Closed” AI Systems

White House Enters Debate on “Open” Versus “Closed” AI Systems

AI and Data Science News posted by ODSC Team Feb 22, 2024 According to the AP, the Biden administration has stepped into the ongoing debate surrounding the openness...

Nvidia’s Multi-Trillion Dollar Market Value Sparks Tech Rally

Nvidia’s Multi-Trillion Dollar Market Value Sparks Tech Rally

AI and Data Science News posted by ODSC Team Feb 22, 2024 In an unprecedented surge, Nvidia’s market capitalization galloped past the $2 trillion mark post-Thursday’s trading session,...

podcast square

EML

37 Research Topics In Data Science To Stay On Top Of

As a data scientist, staying on top of the latest research in your field is essential.

The data science landscape changes rapidly, and new techniques and tools are constantly being developed.

To keep up with the competition, you need to be aware of the latest trends and topics in data science research.

In this article, we will provide an overview of 37 hot research topics in data science.

We will discuss each topic in detail, including its significance and potential applications.

These topics could be an idea for a thesis or simply topics you can research independently.

Stay tuned – this is one blog post you don’t want to miss!

37 Research Topics in Data Science

1.) predictive modeling.

Predictive modeling is a significant portion of data science and a topic you must be aware of.

Simply put, it is the process of using historical data to build models that can predict future outcomes.

Predictive modeling has many applications, from marketing and sales to financial forecasting and risk management.

As businesses increasingly rely on data to make decisions, predictive modeling is becoming more and more important.

While it can be complex, predictive modeling is a powerful tool that gives businesses a competitive advantage.

predictive modeling

2.) Big Data Analytics

These days, it seems like everyone is talking about big data.

And with good reason – organizations of all sizes are sitting on mountains of data, and they’re increasingly turning to data scientists to help them make sense of it all.

But what exactly is big data? And what does it mean for data science?

Simply put, big data is a term used to describe datasets that are too large and complex for traditional data processing techniques.

Big data typically refers to datasets of a few terabytes or more.

But size isn’t the only defining characteristic – big data is also characterized by its high Velocity (the speed at which data is generated), Variety (the different types of data), and Volume (the amount of the information).

Given the enormity of big data, it’s not surprising that organizations are struggling to make sense of it all.

That’s where data science comes in.

Data scientists use various methods to wrangle big data, including distributed computing and other decentralized technologies.

With the help of data science, organizations are beginning to unlock the hidden value in their big data.

By harnessing the power of big data analytics, they can improve their decision-making, better understand their customers, and develop new products and services.

3.) Auto Machine Learning

Auto machine learning is a research topic in data science concerned with developing algorithms that can automatically learn from data without intervention.

This area of research is vital because it allows data scientists to automate the process of writing code for every dataset.

This allows us to focus on other tasks, such as model selection and validation.

Auto machine learning algorithms can learn from data in a hands-off way for the data scientist – while still providing incredible insights.

This makes them a valuable tool for data scientists who either don’t have the skills to do their own analysis or are struggling.

Auto Machine Learning

4.) Text Mining

Text mining is a research topic in data science that deals with text data extraction.

This area of research is important because it allows us to get as much information as possible from the vast amount of text data available today.

Text mining techniques can extract information from text data, such as keywords, sentiments, and relationships.

This information can be used for various purposes, such as model building and predictive analytics.

5.) Natural Language Processing

Natural language processing is a data science research topic that analyzes human language data.

This area of research is important because it allows us to understand and make sense of the vast amount of text data available today.

Natural language processing techniques can build predictive and interactive models from any language data.

Natural Language processing is pretty broad, and recent advances like GPT-3 have pushed this topic to the forefront.

natural language processing

6.) Recommender Systems

Recommender systems are an exciting topic in data science because they allow us to make better products, services, and content recommendations.

Businesses can better understand their customers and their needs by using recommender systems.

This, in turn, allows them to develop better products and services that meet the needs of their customers.

Recommender systems are also used to recommend content to users.

This can be done on an individual level or at a group level.

Think about Netflix, for example, always knowing what you want to watch!

Recommender systems are a valuable tool for businesses and users alike.

7.) Deep Learning

Deep learning is a research topic in data science that deals with artificial neural networks.

These networks are composed of multiple layers, and each layer is formed from various nodes.

Deep learning networks can learn from data similarly to how humans learn, irrespective of the data distribution.

This makes them a valuable tool for data scientists looking to build models that can learn from data independently.

The deep learning network has become very popular in recent years because of its ability to achieve state-of-the-art results on various tasks.

There seems to be a new SOTA deep learning algorithm research paper on  https://arxiv.org/  every single day!

deep learning

8.) Reinforcement Learning

Reinforcement learning is a research topic in data science that deals with algorithms that can learn on multiple levels from interactions with their environment.

This area of research is essential because it allows us to develop algorithms that can learn non-greedy approaches to decision-making, allowing businesses and companies to win in the long term compared to the short.

9.) Data Visualization

Data visualization is an excellent research topic in data science because it allows us to see our data in a way that is easy to understand.

Data visualization techniques can be used to create charts, graphs, and other visual representations of data.

This allows us to see the patterns and trends hidden in our data.

Data visualization is also used to communicate results to others.

This allows us to share our findings with others in a way that is easy to understand.

There are many ways to contribute to and learn about data visualization.

Some ways include attending conferences, reading papers, and contributing to open-source projects.

data visualization

10.) Predictive Maintenance

Predictive maintenance is a hot topic in data science because it allows us to prevent failures before they happen.

This is done using data analytics to predict when a failure will occur.

This allows us to take corrective action before the failure actually happens.

While this sounds simple, avoiding false positives while keeping recall is challenging and an area wide open for advancement.

11.) Financial Analysis

Financial analysis is an older topic that has been around for a while but is still a great field where contributions can be felt.

Current researchers are focused on analyzing macroeconomic data to make better financial decisions.

This is done by analyzing the data to identify trends and patterns.

Financial analysts can use this information to make informed decisions about where to invest their money.

Financial analysis is also used to predict future economic trends.

This allows businesses and individuals to prepare for potential financial hardships and enable companies to be cash-heavy during good economic conditions.

Overall, financial analysis is a valuable tool for anyone looking to make better financial decisions.

Financial Analysis

12.) Image Recognition

Image recognition is one of the hottest topics in data science because it allows us to identify objects in images.

This is done using artificial intelligence algorithms that can learn from data and understand what objects you’re looking for.

This allows us to build models that can accurately recognize objects in images and video.

This is a valuable tool for businesses and individuals who want to be able to identify objects in images.

Think about security, identification, routing, traffic, etc.

Image Recognition has gained a ton of momentum recently – for a good reason.

13.) Fraud Detection

Fraud detection is a great topic in data science because it allows us to identify fraudulent activity before it happens.

This is done by analyzing data to look for patterns and trends that may be associated with the fraud.

Once our machine learning model recognizes some of these patterns in real time, it immediately detects fraud.

This allows us to take corrective action before the fraud actually happens.

Fraud detection is a valuable tool for anyone who wants to protect themselves from potential fraudulent activity.

fraud detection

14.) Web Scraping

Web scraping is a controversial topic in data science because it allows us to collect data from the web, which is usually data you do not own.

This is done by extracting data from websites using scraping tools that are usually custom-programmed.

This allows us to collect data that would otherwise be inaccessible.

For obvious reasons, web scraping is a unique tool – giving you data your competitors would have no chance of getting.

I think there is an excellent opportunity to create new and innovative ways to make scraping accessible for everyone, not just those who understand Selenium and Beautiful Soup.

15.) Social Media Analysis

Social media analysis is not new; many people have already created exciting and innovative algorithms to study this.

However, it is still a great data science research topic because it allows us to understand how people interact on social media.

This is done by analyzing data from social media platforms to look for insights, bots, and recent societal trends.

Once we understand these practices, we can use this information to improve our marketing efforts.

For example, if we know that a particular demographic prefers a specific type of content, we can create more content that appeals to them.

Social media analysis is also used to understand how people interact with brands on social media.

This allows businesses to understand better what their customers want and need.

Overall, social media analysis is valuable for anyone who wants to improve their marketing efforts or understand how customers interact with brands.

social media

16.) GPU Computing

GPU computing is a fun new research topic in data science because it allows us to process data much faster than traditional CPUs .

Due to how GPUs are made, they’re incredibly proficient at intense matrix operations, outperforming traditional CPUs by very high margins.

While the computation is fast, the coding is still tricky.

There is an excellent research opportunity to bring these innovations to non-traditional modules, allowing data science to take advantage of GPU computing outside of deep learning.

17.) Quantum Computing

Quantum computing is a new research topic in data science and physics because it allows us to process data much faster than traditional computers.

It also opens the door to new types of data.

There are just some problems that can’t be solved utilizing outside of the classical computer.

For example, if you wanted to understand how a single atom moved around, a classical computer couldn’t handle this problem.

You’ll need to utilize a quantum computer to handle quantum mechanics problems.

This may be the “hottest” research topic on the planet right now, with some of the top researchers in computer science and physics worldwide working on it.

You could be too.

quantum computing

18.) Genomics

Genomics may be the only research topic that can compete with quantum computing regarding the “number of top researchers working on it.”

Genomics is a fantastic intersection of data science because it allows us to understand how genes work.

This is done by sequencing the DNA of different organisms to look for insights into our and other species.

Once we understand these patterns, we can use this information to improve our understanding of diseases and create new and innovative treatments for them.

Genomics is also used to study the evolution of different species.

Genomics is the future and a field begging for new and exciting research professionals to take it to the next step.

19.) Location-based services

Location-based services are an old and time-tested research topic in data science.

Since GPS and 4g cell phone reception became a thing, we’ve been trying to stay informed about how humans interact with their environment.

This is done by analyzing data from GPS tracking devices, cell phone towers, and Wi-Fi routers to look for insights into how humans interact.

Once we understand these practices, we can use this information to improve our geotargeting efforts, improve maps, find faster routes, and improve cohesion throughout a community.

Location-based services are used to understand the user, something every business could always use a little bit more of.

While a seemingly “stale” field, location-based services have seen a revival period with self-driving cars.

GPS

20.) Smart City Applications

Smart city applications are all the rage in data science research right now.

By harnessing the power of data, cities can become more efficient and sustainable.

But what exactly are smart city applications?

In short, they are systems that use data to improve city infrastructure and services.

This can include anything from traffic management and energy use to waste management and public safety.

Data is collected from various sources, including sensors, cameras, and social media.

It is then analyzed to identify tendencies and habits.

This information can make predictions about future needs and optimize city resources.

As more and more cities strive to become “smart,” the demand for data scientists with expertise in smart city applications is only growing.

21.) Internet Of Things (IoT)

The Internet of Things, or IoT, is exciting and new data science and sustainability research topic.

IoT is a network of physical objects embedded with sensors and connected to the internet.

These objects can include everything from alarm clocks to refrigerators; they’re all connected to the internet.

That means that they can share data with computers.

And that’s where data science comes in.

Data scientists are using IoT data to learn everything from how people use energy to how traffic flows through a city.

They’re also using IoT data to predict when an appliance will break down or when a road will be congested.

Really, the possibilities are endless.

With such a wide-open field, it’s easy to see why IoT is being researched by some of the top professionals in the world.

internet of things

22.) Cybersecurity

Cybersecurity is a relatively new research topic in data science and in general, but it’s already garnering a lot of attention from businesses and organizations.

After all, with the increasing number of cyber attacks in recent years, it’s clear that we need to find better ways to protect our data.

While most of cybersecurity focuses on infrastructure, data scientists can leverage historical events to find potential exploits to protect their companies.

Sometimes, looking at a problem from a different angle helps, and that’s what data science brings to cybersecurity.

Also, data science can help to develop new security technologies and protocols.

As a result, cybersecurity is a crucial data science research area and one that will only become more important in the years to come.

23.) Blockchain

Blockchain is an incredible new research topic in data science for several reasons.

First, it is a distributed database technology that enables secure, transparent, and tamper-proof transactions.

Did someone say transmitting data?

This makes it an ideal platform for tracking data and transactions in various industries.

Second, blockchain is powered by cryptography, which not only makes it highly secure – but is a familiar foe for data scientists.

Finally, blockchain is still in its early stages of development, so there is much room for research and innovation.

As a result, blockchain is a great new research topic in data science that vows to revolutionize how we store, transmit and manage data.

blockchain

24.) Sustainability

Sustainability is a relatively new research topic in data science, but it is gaining traction quickly.

To keep up with this demand, The Wharton School of the University of Pennsylvania has  started to offer an MBA in Sustainability .

This demand isn’t shocking, and some of the reasons include the following:

Sustainability is an important issue that is relevant to everyone.

Datasets on sustainability are constantly growing and changing, making it an exciting challenge for data scientists.

There hasn’t been a “set way” to approach sustainability from a data perspective, making it an excellent opportunity for interdisciplinary research.

As data science grows, sustainability will likely become an increasingly important research topic.

25.) Educational Data

Education has always been a great topic for research, and with the advent of big data, educational data has become an even richer source of information.

By studying educational data, researchers can gain insights into how students learn, what motivates them, and what barriers these students may face.

Besides, data science can be used to develop educational interventions tailored to individual students’ needs.

Imagine being the researcher that helps that high schooler pass mathematics; what an incredible feeling.

With the increasing availability of educational data, data science has enormous potential to improve the quality of education.

online education

26.) Politics

As data science continues to evolve, so does the scope of its applications.

Originally used primarily for business intelligence and marketing, data science is now applied to various fields, including politics.

By analyzing large data sets, political scientists (data scientists with a cooler name) can gain valuable insights into voting patterns, campaign strategies, and more.

Further, data science can be used to forecast election results and understand the effects of political events on public opinion.

With the wealth of data available, there is no shortage of research opportunities in this field.

As data science evolves, so does our understanding of politics and its role in our world.

27.) Cloud Technologies

Cloud technologies are a great research topic.

It allows for the outsourcing and sharing of computer resources and applications all over the internet.

This lets organizations save money on hardware and maintenance costs while providing employees access to the latest and greatest software and applications.

I believe there is an argument that AWS could be the greatest and most technologically advanced business ever built (Yes, I know it’s only part of the company).

Besides, cloud technologies can help improve team members’ collaboration by allowing them to share files and work on projects together in real-time.

As more businesses adopt cloud technologies, data scientists must stay up-to-date on the latest trends in this area.

By researching cloud technologies, data scientists can help organizations to make the most of this new and exciting technology.

cloud technologies

28.) Robotics

Robotics has recently become a household name, and it’s for a good reason.

First, robotics deals with controlling and planning physical systems, an inherently complex problem.

Second, robotics requires various sensors and actuators to interact with the world, making it an ideal application for machine learning techniques.

Finally, robotics is an interdisciplinary field that draws on various disciplines, such as computer science, mechanical engineering, and electrical engineering.

As a result, robotics is a rich source of research problems for data scientists.

29.) HealthCare

Healthcare is an industry that is ripe for data-driven innovation.

Hospitals, clinics, and health insurance companies generate a tremendous amount of data daily.

This data can be used to improve the quality of care and outcomes for patients.

This is perfect timing, as the healthcare industry is undergoing a significant shift towards value-based care, which means there is a greater need than ever for data-driven decision-making.

As a result, healthcare is an exciting new research topic for data scientists.

There are many different ways in which data can be used to improve healthcare, and there is a ton of room for newcomers to make discoveries.

healthcare

30.) Remote Work

There’s no doubt that remote work is on the rise.

In today’s global economy, more and more businesses are allowing their employees to work from home or anywhere else they can get a stable internet connection.

But what does this mean for data science? Well, for one thing, it opens up a whole new field of research.

For example, how does remote work impact employee productivity?

What are the best ways to manage and collaborate on data science projects when team members are spread across the globe?

And what are the cybersecurity risks associated with working remotely?

These are just a few of the questions that data scientists will be able to answer with further research.

So if you’re looking for a new topic to sink your teeth into, remote work in data science is a great option.

31.) Data-Driven Journalism

Data-driven journalism is an exciting new field of research that combines the best of both worlds: the rigor of data science with the creativity of journalism.

By applying data analytics to large datasets, journalists can uncover stories that would otherwise be hidden.

And telling these stories compellingly can help people better understand the world around them.

Data-driven journalism is still in its infancy, but it has already had a major impact on how news is reported.

In the future, it will only become more important as data becomes increasingly fluid among journalists.

It is an exciting new topic and research field for data scientists to explore.

journalism

32.) Data Engineering

Data engineering is a staple in data science, focusing on efficiently managing data.

Data engineers are responsible for developing and maintaining the systems that collect, process, and store data.

In recent years, there has been an increasing demand for data engineers as the volume of data generated by businesses and organizations has grown exponentially.

Data engineers must be able to design and implement efficient data-processing pipelines and have the skills to optimize and troubleshoot existing systems.

If you are looking for a challenging research topic that would immediately impact you worldwide, then improving or innovating a new approach in data engineering would be a good start.

33.) Data Curation

Data curation has been a hot topic in the data science community for some time now.

Curating data involves organizing, managing, and preserving data so researchers can use it.

Data curation can help to ensure that data is accurate, reliable, and accessible.

It can also help to prevent research duplication and to facilitate the sharing of data between researchers.

Data curation is a vital part of data science. In recent years, there has been an increasing focus on data curation, as it has become clear that it is essential for ensuring data quality.

As a result, data curation is now a major research topic in data science.

There are numerous books and articles on the subject, and many universities offer courses on data curation.

Data curation is an integral part of data science and will only become more important in the future.

businessman

34.) Meta-Learning

Meta-learning is gaining a ton of steam in data science. It’s learning how to learn.

So, if you can learn how to learn, you can learn anything much faster.

Meta-learning is mainly used in deep learning, as applications outside of this are generally pretty hard.

In deep learning, many parameters need to be tuned for a good model, and there’s usually a lot of data.

You can save time and effort if you can automatically and quickly do this tuning.

In machine learning, meta-learning can improve models’ performance by sharing knowledge between different models.

For example, if you have a bunch of different models that all solve the same problem, then you can use meta-learning to share the knowledge between them to improve the cluster (groups) overall performance.

I don’t know how anyone looking for a research topic could stay away from this field; it’s what the  Terminator  warned us about!

35.) Data Warehousing

A data warehouse is a system used for data analysis and reporting.

It is a central data repository created by combining data from multiple sources.

Data warehouses are often used to store historical data, such as sales data, financial data, and customer data.

This data type can be used to create reports and perform statistical analysis.

Data warehouses also store data that the organization is not currently using.

This type of data can be used for future research projects.

Data warehousing is an incredible research topic in data science because it offers a variety of benefits.

Data warehouses help organizations to save time and money by reducing the need for manual data entry.

They also help to improve the accuracy of reports and provide a complete picture of the organization’s performance.

Data warehousing feels like one of the weakest parts of the Data Science Technology Stack; if you want a research topic that could have a monumental impact – data warehousing is an excellent place to look.

data warehousing

36.) Business Intelligence

Business intelligence aims to collect, process, and analyze data to help businesses make better decisions.

Business intelligence can improve marketing, sales, customer service, and operations.

It can also be used to identify new business opportunities and track competition.

BI is business and another tool in your company’s toolbox to continue dominating your area.

Data science is the perfect tool for business intelligence because it combines statistics, computer science, and machine learning.

Data scientists can use business intelligence to answer questions like, “What are our customers buying?” or “What are our competitors doing?” or “How can we increase sales?”

Business intelligence is a great way to improve your business’s bottom line and an excellent opportunity to dive deep into a well-respected research topic.

37.) Crowdsourcing

One of the newest areas of research in data science is crowdsourcing.

Crowdsourcing is a process of sourcing tasks or projects to a large group of people, typically via the internet.

This can be done for various purposes, such as gathering data, developing new algorithms, or even just for fun (think: online quizzes and surveys).

But what makes crowdsourcing so powerful is that it allows businesses and organizations to tap into a vast pool of talent and resources they wouldn’t otherwise have access to.

And with the rise of social media, it’s easier than ever to connect with potential crowdsource workers worldwide.

Imagine if you could effect that, finding innovative ways to improve how people work together.

That would have a huge effect.

crowd sourcing

Final Thoughts, Are These Research Topics In Data Science For You?

Thirty-seven different research topics in data science are a lot to take in, but we hope you found a research topic that interests you.

If not, don’t worry – there are plenty of other great topics to explore.

The important thing is to get started with your research and find ways to apply what you learn to real-world problems.

We wish you the best of luck as you begin your data science journey!

Other Data Science Articles

We love talking about data science; here are a couple of our favorite articles:

  • Why Are You Interested In Data Science?
  • Recent Posts

Dylan Kaplan

  • Jenkins pipeline vs. GitLab pipeline [With Example Code] - February 22, 2024
  • GitLab CI/CD PyTest Tutorial for Beginners [WITH CODE EXAMPLE] - February 22, 2024
  • MLOps vs Data Engineer [Which Will You Like More?] - February 22, 2024

863-225-3925

4131 Dolphin Dr Unit 81315, Tampa, FL 33617

© 2024 enjoymachinelearning.com

Thesis Topics

This list includes topics for potential bachelor or master theses, guided research, projects, seminars, and other activities. Search with Ctrl+F for desired keywords, e.g. ‘machine learning’ or others.

PLEASE NOTE: If you are interested in any of these topics, click the respective supervisor link to send a message with a simple CV, grade sheet, and topic ideas (if any). We will answer shortly.

Of course, your own ideas are always welcome!

Generating images for training Image Super-Resolution models

Type of work:.

  • Guided Research
  • deep learning
  • single image super-resolution
  • syntethic datasets / dataset generation

Description:

Typically, Single Image Super-Resolution (SISR) models train on expressive real images (e.g., DIV2K and/or Flickr2K). This work aims to rethink the need of real images for training SISR models. In other words: do we need real images to learn useful upscaling mappings? For that, the proposed work should investigate different methods for generating artificial datasets that might be suitable for SISR models, see [2]. The resulting models trained on the artifically generated training sets should then be evaluated on real test datasets (Set5, Set14, BSDS100, …) and analyze its outcomes.

  • [1] Hitchhiker’s Guide to Super-Resolution: Introduction and Recent Advances
  • [2] Learning to See by Looking at Noise

Machine Learning-based Surrogate Models for Accelerated Flow Simulations

  • Machine Learning
  • Microstructure Property Prediction
  • Surrogate Modeling

Surrogate modeling involves creating a simplified and computationally efficient machine learning model that approximates the behavior of a complex system, enabling faster predictions and analysis. For complex systems such as fluids, their behavior is governed by partial differential equations. By solving these PDEs, one can predict how a fluid behaves in a specific environment and conditions. The computational time and resources needed to solve a PDE system depend on the size of the fluid domain and the complexity of the PDE. In practical applications where multiple environments and conditions are to be studied, it becomes very expensive to generate many solutions to such PDEs. Here, modern machine learning or deep learning-based surrogate models which offer fast inference times in the online phase are of interest.

In this work, the focus will be on developing surrogate models to replace the flow simulations in fiber-reinforced composite materials governed by the Navier-Stokes equation. Using a conventional PDE solver, a dataset of reference solutions was generated for supervised learning. In this thesis, your tasks will include the conceptualization and implementation of different ML architectures suited for this task, training and evaluation of the models on the available dataset. You will start with simple fully connected architectures and later extend it to 3D convolutional architectures. Also of interest is the infusion of the available domain knowledge into the ML models, known as physics-informed machine learning.

By applying ML to fluid applications, you will learn to acquire the right amount of domain specific knowledge and analyze your results together with domain experts from the field.

If you are interested, please send me an email with your Curriculum Vitae (CV), your Transcript of records and a short statement about your background in related topics.

References:

  • Santos, J.E., Xu, D., Jo, H., Landry, C.J., Prodanović, M., Pyrcz, M.J., 2020. PoreFlow-Net: A 3D convolutional neural network to predict fluid flow through porous media. Advances in Water Resources 138, 103539. https://doi.org/10.1016/j.advwatres.2020.103539
  • Kashefi, A., Mukerji, T., 2021. Point-cloud deep learning of porous media for permeability prediction. Physics of Fluids 33, 097109. https://doi.org/10.1063/5.0063904

Segmentation of Shoe Trace Images

  • benchmarking
  • image segmentation
  • keypoint extraction
  • self-attention

Help fight crime with AI! The DFKI and the Artificial Intelligence Transferlab of the State Criminal Police Office (Landskriminalamt) are searching for master candidates eager to apply their knowledge in AI to support crime scene analysis. The student will have the opportunity to visit the Transferlab in Mainz for an in-depth introduction to the topic and full access to DFKI’s computing cluster infrastructure.

General goal: improve identification of specific markers normally present in shoe trace images acquired in crime scenes.

Specific goals:

  • [benchmarking] evaluate existing image segmentation models in the context of shoe trace analysis;
  • [research] propose a segmentation model combining semantics and keypoint information tailored to specific markers present in crime scene photographs;
  • [research] assess model performance on labeled data.
  • [research] definition of limits and requirements for the existing training- and test-data.

Retrieval of Shoe Sole Images

  • graph neural networks
  • image retrieval

General goal: improve retrieval of shoe sole images acquired in laboratory, i.e. under controlled conditions and used as reference by forensics specialists.

  • [benchmarking] evaluate existing image retrieval approaches in the context of shoe trace recognition;
  • [research] propose a graph network architecture based on keypoint information extracted from the images.
  • [research] evaluate performance of proposed model against existing methods.

Sherlock Holmes goes AI - Generative comics art of detective scenes and identikits

  • Bias in image generation models
  • Deep Learning Frameworks
  • Frontend visualization
  • Speech-To-Text, Text-to-Image Models
  • Transformers, Diffusion Models, Hugging Face

Sherlock Holmes is taking the statement of the witness. The witness is describing the appearance of the perpetrator and the forensic setting they still remember. Your task as the AI investigator will be to generate a comic sketch of the scene and phantom images of the accused person based on the spoken statement of the witness. For this you will use state-of-the-art transformers and visualize the output in an application. As AI investigator you will detect, qualify and quantify bias in the images which are produced by different generation models you have chosen.

This work is embedded in the DFKI KI4Pol lab together with the law enforcement agencies. The stories are fictional you will not work on true crime.

Requirements:

  • German level B1/2 or equivalent
  • Outstanding academic achievements
  • Motivational cover letter

Generative Adversarial Networks for Agricultural Yield Prediction

  • Deep Learning
  • Generative Adversarial Networks
  • Yield Prediction

Agricultural yield prediction has been an essential research area for many years, as it helps farmers and policymakers to make informed decisions about crop management, resource allocation, and food security. Computer vision and machine learning techniques have shown promising results in predicting crop yield, but there is still room for improvement in the accuracy and precision of these predictions. Generative Adversarial Networks (GANs) are a type of neural network that has shown success in generating realistic images, which can be leveraged for the prediction of agricultural yields.

  • ‘Goodfellow, Ian, et al. “Generative adversarial networks.” Communications of the ACM 63.11 (2020)': 139-144.
  • ‘Z. Xu, J. Du, J. Wang, C. Jiang and Y. Ren, “Satellite Image Prediction Relying on GAN and LSTM Neural Networks,” ICC 2019 - 2019 IEEE International Conference on Communications (ICC), Shanghai, China, 2019, pp. 1-6, doi’: 10.1109/ICC.2019.8761462.
  • ‘Drees, Lukas, et al. “Temporal prediction and evaluation of brassica growth in the field using conditional generative adversarial networks.” Computers and Electronics in Agriculture 190 (2021)': 106415

Knowledge Graphs für das Immobilienmanagement

  • corporate memory
  • knowledge graph

Das Management von Immobilien ist komplex und umfasst verschiedenste Informationsquellen und -objekte zur Durchführung der Prozesse. Ein Corporate Memory kann hier unterstützen in der Analyse und Abbildung des Informationsraums um Wissensdienste zu ermöglichen. Aufgabe ist es, eine Ontologie für das Immobilienmanagement zu entwerfen und beispielhaft ein Szenario zu entwickeln. Für die Materialien und Anwendungspartner sind gute Deutschkenntnisse erforderlich.

Fault and Efficiency Prediction in High Performance Computing

  • Master Thesis
  • event data modelling
  • survival modelling
  • time series

High use of resources are thought to be an indirect cause of failures in large cluster systems, but little work has systematically investigated the role of high resource usage on system failures, largely due to the lack of a comprehensive resource monitoring tool which resolves resource use by job and node. This project studies log data of the DFKI Kaiserslautern high performance cluster to consider the predictability of adverse events (node failure, GPU freeze), energy usage and identify the most relevant data within. The second supervisor for this work is Joachim Folz.

Data is available via Prometheus -compatible system:

  • Node exporter
  • DCGM exporter
  • Slurm exporter
  • Linking Resource Usage Anomalies with System Failures from Cluster Log Data
  • Deep Survival Models

Feel free to reach out if the topic sounds interesting or if you have ideas related to this work. We can then brainstorm a specific research question together. Link to my personal website.

Construction & Application of Enterprise Knowledge Graphs in the E-Invoicing Domain

  • Guided Research Project
  • knowledge graphs
  • knowledge services
  • linked data
  • semantic web

In recent years knowledge graphs received a lot of attention as well in industry as in science. Knowledge graphs consist of entities and relationships between them and allow integrating new knowledge arbitrarily. Famous instances in industry are knowledge graphs by Microsoft, Google, Facebook or IBM. But beyond these ones, knowledge graphs are also adopted in more domain specific scenarios such as in e-Procurement, e-Invoicing and purchase-to-pay processes. The objective in theses and projects is to explore particular aspects of constructing and/or applying knowledge graphs in the domain of purchase-to-pay processes and e-Invoicing.

Learning Analytics in Education

  • affective state
  • cognitive state
  • machine learning

Anomaly detection in time-series

  • explainability

Working on deep neural networks for making the time-series anomaly detection process more robust. An important aspect of this process is explainability of the decision taken by a network.

Time Series Forecasting Using transformer Networks

  • time series forecasting
  • transformer networks

Transformer networks have emerged as competent architecture for modeling sequences. This research will primarily focus on using transformer networks for forecasting time series (multivariate/ univariate) and may also involve fusing knowledge into the machine learning architecture.

On This Page

  • Machine Learning Tutorial
  • Data Analysis Tutorial
  • Python - Data visualization tutorial
  • Machine Learning Projects
  • Machine Learning Interview Questions
  • Machine Learning Mathematics
  • Deep Learning Tutorial
  • Deep Learning Project
  • Deep Learning Interview Questions
  • Computer Vision Tutorial
  • Computer Vision Projects
  • NLP Project
  • NLP Interview Questions
  • Statistics with Python
  • 100 Days of Machine Learning

Related Articles

  • Solve Coding Problems
  • How Amazon Uses Machine Learning?
  • What Are DeepFakes And How Dangerous Are They?
  • 10 Most Interesting Chatbots in the World
  • 5 Algorithms that Demonstrate Artificial Intelligence Bias
  • Impact of AI and ML On Warfare Techniques
  • What is the Role of Artificial Intelligence in Fighting Coronavirus?
  • What is Artificial Intelligence as a Service (AIaaS) in the Tech Industry?
  • Top 10 Business Intelligence Platforms in 2020
  • Artificial Intelligence Could be a Better Doctor
  • Top 7 Artificial Intelligence and Machine Learning Trends For 2022
  • What is IBM Watson and Its Services?
  • 10 Best Artificial Intelligence Project Ideas To Kick-Start Your Career
  • Is AI Really a Threat to Cybersecurity?
  • 5 Best Humanoid Robots in The World
  • Can Artificial Intelligence Help in Curing Cancer?
  • 8 Best Artificial Intelligence Books For Beginners in 2023
  • AI | Phrase and Grammar structure in Natural Language
  • What is Artificial General Intelligence (AGI)?
  • Linguistic variable And Linguistic hedges

8 Best Topics for Research and Thesis in Artificial Intelligence

Best-Topics-for-Research-and-Thesis-in-Artificial-Intelligence

1. Machine Learning

2. deep learning, 3. reinforcement learning, 4. robotics, 5. natural language processing, 6. computer vision, 7. recommender systems, 8. internet of things, please login to comment....

author

  • Artificial Intelligence
  • Machine Learning

Improve your Coding Skills with Practice

 alt=

What kind of Experience do you want to share?

  • Faculty of Arts and Sciences
  • FAS Theses and Dissertations
  • Communities & Collections
  • By Issue Date
  • FAS Department
  • Quick submit
  • Waiver Generator
  • DASH Stories
  • Accessibility
  • COVID-related Research

Terms of Use

  • Privacy Policy
  • By Collections
  • By Departments

Undergraduate Fundamentals of Machine Learning

Thumbnail

Citable link to this page

Collections.

  • FAS Theses and Dissertations [5858]

Contact administrator regarding this item (to report mistakes or request changes)

  • Studienangebot
  • Vorlesungsverzeichnis
  • Bewerbung & Zulassung
  • Schnupperstudium
  • Gasthörerstudium
  • Beratung und Information
  • Profil Universität Hildesheim
  • Campusleben
  • Studienfinanzierung
  • Studienbeiträge
  • Zentrale Studienberatung
  • International Office
  • Studieren mit Familie
  • Studieren mit Behinderung
  • Termine & Fristen
  • Anfahrts- und Lageplan
  • Fachbereiche
  • Vorlesungsverzeichnis LSF
  • Studien- & Prüf.-Ordnungen
  • Einschreibung, Rückmeldung, Prüfungen...
  • Universitätsbibliothek
  • Rechenzentrum
  • Studierenden-Vertretung
  • Jobs und Praktika
  • Information & Beratung
  • Gleichstellungsbüro
  • Konferenz wiss. Mitarbeiterinnen
  • Personalrat
  • Rat & Hilfe bei Suchtfragen
  • Schwerbehindertenvertretung
  • Lokale Angebote
  • Übergreifende Angebote
  • Gremientermine
  • Elektr. Verkündungsblatt
  • Arbeiten mit Familie
  • Veranstaltungen
  • Neuigkeiten
  • UniMagazin/UniJournal
  • Stiftung Uni Hildesheim
  • Fundraising
  • Universitätsgesellschaft
  • Ehrenmitglieder
  • Alumni-Vereine
  • Gasthörer-Studium
  • Stellenmarkt
  • Universitätsverlag
  • Universitätsarchiv
  • Pressestelle
  • Anfahrts- & Lageplan

Logo Stiftung Universität Hildesheim

  • Degree Programs
  • Bachelor/Master
  • Publications
  • ISMLL Cluster
  • Travel Information
  • Legal notice

Dekobild im Seitenkopf ISMLL

  • University of Hildesheim »
  • Department of Mathematics, Natural Science, Economics and Computer Science »
  • Institute of Computer Science »
  • Information Systems and Machine Learning Lab (ISMLL)

Master Thesis at ISMLL

  • How to Start a Master Thesis
  • Available Thesis Topics
  • Current Master Theses
  • Past Master Theses

Available Supervisors

Available master thesis topics at ismll.

Go is an old two-player board game with origin in Asia. The game is easy to learn but hard to master which makes it so popular all over the world. Go is also interesting for artificial intelligence research. For Chess already programs exist that can beat any human player. For Go the strongest AI players are far away from beating human experts. There are two main reasons: i) the search space of Go is extremely large since many move options exist and ii) moves can have an important long-term effect such that no good heuristic was found that can evaluate a board state. Thus, techniques that led to success in games like Chess, cannot be applied successfully in Go.

Strong state of the art Go computers are using a combination of Monte Carlo Tree Search (a heuristic search algorithm) and move prediction algorithms that are pruning the search space and guide the search. The aim of this thesis is to compare these move prediction algorithms. For that reason, two algorithms need to be integrated into an existing Go computer (e.g. Fuego ) and the better move predictor will be estimated by a direct contest.

Learning the structure of Bayesian Networks from data is one of the core tasks in Bayesian Networks that has been addressed by several algorithms recently. Starting from the CGNM/BN implementation in Java, that supports basic tasks as IO, inferencing and parameter learning, first some simple and fast algorithms like K2 should be implemented. Building on that, two more complex algorithms (PC and GES) should be implemented efficienlty. The performance of these algorithms should be evaluated in terms of quality of the solution found as well as runtime on real life datasets. Finally, synthetic datasets created from synthetic BNs should be used to assess the algorithms under controllable conditions.

Many web-scale applications captures data that are multi-way and are represented through a multidimensional arrays or Tensor. Example of such applications are Social network analysis (user, user and time), Recommender Systems (user, product, time and auxiliary relations), and Topic modeling (topic, document, word). Tensor factorization is used as a tool to learn latent features from the mutli-way dataset. The scalability of these Tensor Factorization is an important aspect, as these tensor easily exhaust the memory of a single node. The existing methods partitions the data among workers, but usually, require to hold latent matrices in memory.

In this thesis you have to explore data partitioning options, such that each worker should not load the complete latent matrices. You have to investigate the scalability scalability in terms of the datasizes, the number of nodes, and the convergence behavior

  • 1. Shin, Kijung, Lee Sael, and U. Kang. "Fully scalable methods for distributed tensor factorization." IEEE Transactions on Knowledge and Data Engineering, 2017.
  • 2. Beutel, Alex, et al. "Flexifact: Scalable flexible factorization of coupled tensors on hadoop." Proceedings of the 2014 SIAM International Conference on Data Mining. Society for Industrial and Applied Mathematics, 2014.
  • 3. Oh, Jinoh, et al. "S-hot: Scalable high-order tucker decomposition." Proceedings of the Tenth ACM International Conference on Web Search and Data Mining. ACM, 2017.

The reinforcement learning (RL) has seen wide up-take by the research community as well as the industry. The RL setting consists of an agent which interacts with the environment and learns a policy that is optimal to solve a certain problem. If the environment is complex the learning process could slow down, requiring significantly amount of resources to reach the optimal solution. In this direction the speedup of the learning process could be achieved by using parallel and distributed concepts as well as enhancing the exploration capacity of the agent.

In this thesis you have to develop a scalable RL framework that exploits the parallelism to speed up the learning process and diversify the exploration capacity of the agents.

  • 1. Dimakopoulou, Maria and Van Roy, Benjamin. "Coordinated Exploration in Concurrent Reinforcement Learning" Proceedings of the 35th International Conference on Machine Learning, 2018.
  • 2. Volodymyr Mnih, et al. "Asynchronous Methods for Deep Reinforcement Learning" Proceedings of The 33rd International Conference on Machine Learning, 2016

The big data is generally attributed with the large number of data instances and features. This increase in the size of data is attributed to sophisticated means of capturing data via sensors or growth in online retail services. In both cases the size of data ranges between terabytes to petabytes with billions of features. However, the data features that significantly contribute towards learning are very sparse. The sparse machine learning models are typically used i.e. LASSO. The existing sparse models that work in distributed settings usually start with large models and have a high communication cost. The high communication cost dominates the execution time in distributed settings.

In literature, many screening algorithms[2] exist that reduce dimensionality of the model. But these algorithms are applied as a preprocessing step and are sequential. In this thesis you have to explore the existing screening algorithms by implement and evaluating them in a distributed setting. Interesting aspect would be to explore the compactness of these models, its impact on the communication cost and the convergence. A novel aspect would be to design an efficient distributed optimization algorithm, that exploits compactness of the model.

  • 1. Li, Qingyang, et al. "Parallel Lasso Screening for Big Data Optimization." Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 2016.
  • 2. L. E. Ghaoui, V. Viallon, and T. Rabbani. Safe feature elimination for the lasso and sparse supervised learning problems. Pacific Journal of Optimization, (8):667–698, 2012.
  • 3. R. Tibshirani, J. Bien, J. Friedman, T. Hastie, N. Simon, J. Taylor, and R. J. Tibshirani. Strong rules for discarding predictors in lasso-type problems. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 74(2):245–266, 2012.

Learning in large scale distributed networks is a challenging task, because centralization of data is not feasible. Any data change must be reported to the central server, and it might alter the results due to distributed, dynamically changing and unreliable nature of such ad-hoc networks. Therefore, the goal is to develop distributed mining algorithms that are communication efficient, scalable, asynchronous, and robust to user mobility, which achieve accuracy as close as possible to centralized solutions.

One particular approach is to share very compact but very significant data or models among neighboring compute nodes. Idea is to extract least but effective data or models, which reduces the cost of exchange (communication) and increases the prediction accuracy of the neighboring recipient and hence of overall network. In this regard, this Master thesis would exploit emergence of frequent nearest neighbors in high dimensional data. If such data instances emerge as centroids of local clusters, they are expected to carry useful information about this local neighborhood. Hence, exchanging only such instances, can increase overall prediction accuracy while keep communication load under control.

Your task is to understand the problem and then analyze the baseline approaches for data or model exchange in distributed scenarios. In the end, a working implementation and concrete experimental results are expected. Novel additions to the state-of-art will be highly appreciated during evaluation.

  • Miloš Radovanović, Alexandros Nanopoulos, and Mirjana Ivanović. 2009. Nearest neighbors in high-dimensional data: the emergence and influence of hubs. In Proceedings of the 26th Annual International Conference on Machine Learning (ICML '09)
  • Hock Hee Ang, Vivekanand Gopalkrishnan, Steven C. Hoi, and Wee Keong Ng. 2008. Cascade RSVM in Peer-to-Peer Networks. In Proceedings of the 2008 European Conference on Machine Learning and Knowledge Discovery in Databases - Part I (ECML PKDD '08)

Automatic music transcription (AMT) or dictation is a fundamental problem in the Music Information Retrieval (MIR) area, and is described as the process of converting an acoustic signal into symbolic notations or an equivalent representation. With human beings, this skill is often expected to be acquired by music students after a couple of years into their college degree. Nevertheless, current state-of-the-art algorithms in AMT still have a lot more room for improvement [3] ever since [1] first attempted automatic duet transcription over three decades ago and [2] presented the spectral peak picking for piano note detection. Obtaining a standardized parametric representation of music is important for content-based retrieval, and can aid in music analysis. The specific problem targeted by this thesis is polyphonic music transcription, namely multi-pitch estimation, also referred to as multiple-F0 or multi-pitch detection, which may be considered a core problem in automatic transcription. The challenge presented in this problem is the concurrent presence of distinct notes, leading to interference amongst the corresponding harmonics

  • [1] Dixon, Simon. "On the computer recognition of solo piano music." Proceedings of Australasian computer music conference. 2000.
  • [2] Moorer, James A. "On the transcription of musical sound by computer." Computer Music Journal (1977): 32-38.
  • [3] Sigtia, Siddharth, Emmanouil Benetos, and Simon Dixon. "An end-to-end neural network for polyphonic piano music transcription." IEEE/ACM Transactions on Audio, Speech and Language Processing (TASLP) 24.5 (2016): 927-939.

Parking availability prediction is rapidly gaining interest within the community as an operationally cheap approach to identifying empty parking locations. Parking locations accommodate multiple vehicles and are rarely completely occupied which makes it difficult to predict occupied locations without the augmentation of external data, as the data becomes highly imbalanced. Existing forecasting models neither encapsulate the heterogeneous modes/types of parking data, nor can handle sparse measurements. In this thesis, your task will be to develop a sequence-to-sequence learning framework that addresses the issue of occupancy forecasting. Throught this project, you will build on top of existing methods pertaining to multi-step forecasting, and deliver a model that is competitive with what has been recently done.

  • [1] Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le. "Sequence to sequence learning with neural networks." Advances in neural information processing systems. 2014.
  • [2] Lin, Trista, et al. ”A survey of smart parking solutions.” parking 1524 (2017): 9050.: 32-38.
  • [3] Chen, Xiao. ”Parking occupancy prediction and pattern analysis.” Dept. Comput. Sci., Stanford Univ., Stanford, CA, USA, Tech. Rep. CS229-2014 (2014).

Early diagnosis of disease can increase the chances of finding a cure by providing the proper preventive treatment. With the integration of information technology in the medical domain, electronic health data of patients become readily available for processing. In order to benefit from the raw signals,data and meta-data of the observations, it is critical to learn a rich representation of the data and utilizing the obtained feature representations in a machine learning model that aids in the predictive analysis task. The continuous success of neural networks in machine learning is being recently investigated in health care [1] , [2] , through LSTM’s and recurrent neural networks due to the time-dependent nature of the data. To this end, the objective of this thesis is to develop a strong neural network that learns a meaningful representation of the data and is able to predict disease onset with competitive accuracy as compared to existing baselines.

  • [1] Razavian, Narges, Jake Marcus, and David Sontag. "Multi-task prediction of disease onsets from longitudinal laboratory tests." Machine Learning for Healthcare Conference. 2016.
  • [2] Choi, Edward, et al. "Doctor ai: Predicting clinical events via recurrent neural networks." Machine Learning for Healthcare Conference. 2016.

The notion of attention in machine learning approaches proved successful in domain adaptation, time series forecasting, and image classification. However, all attention techniques are designed on an instance level. In this thesis, you will investigate the a new form of attention, by means of dataset meta-feature similarity. The idea in a nutshell is that, the performance of a target model for a given hyper-parameter configuration depends on the similarity between the target dataset itself and the support dataset, as well as the corrsesponding surrogate value of support dataset .

  • [1] Vinyals, O., Blundell, C., Lillicrap, T., and Wierstra, D. (2016). Matching networks for one shot learning. In Advances in neural information processing systems (pp. 3630-3638).
  • [2] Jomaa, H. S., Schmidt-Thieme, L., and Grabocka, J. (2019). Dataset2vec: Learning dataset meta-features. arXiv preprint arXiv:1905.11063.

Dividing the reinforcement learning task into a hierarchy of problems is one of the solutions provided for environment with a large state space. Some of the existing methods focus on decomposing the reward and learning separate value functions per component [1]. Other methods, referred to as Feudal Learning, split the learning agent into modules, namely a master and slave. The master learns abstract actions and propagates them to the worker, which receivs an inherent reward if he comlies with the master, and aims to maximize the discounted reward [2]. In this thesis, the objective is to focus on another aspect of hierarchical learning, Compositonal Q-learning [3]. Simply put, in compositional Q-learning, the agent aims to maximize the reward of elemental tasks, and composite tasks, which are an ordered sequence of elemental tasks. This type of Learning is interesting in atari games as well as for robotic control, both of which have readliy available source code inthe OpenAI repository[4].

  • [1] Van Seijen, Harm, et al. "Hybrid reward architecture for reinforcement learning." Advances in Neural Information Processing Systems. 2017.
  • [2] Vezhnevets, Alexander Sasha, et al. "Feudal networks for hierarchical reinforcement learning." arXiv preprint arXiv:1703.01161 (2017).
  • [3] Tham, Chen K., and Richard W. Prager. "A modular q-learning architecture for manipulator task decomposition." Machine Learning Proceedings 1994. 1994. 309-317.
  • [4] https://github.com/openai/gym

One of the existing challenges in reinforcement learning is exploration. After an agent is trained, it tends to stick to the actions that have proved to maximize its previous rewards, without any incentive to explore other, possibly more rewarding actions. Several approaches have been proposed to boost exploration which inlcude noise injection into the action space[1], noise injection into the parameter space[2], curiosity drive exploration in which curiosity leads to a higher intrinsic reward [3]. In this thesis, you will focus on uncertainty-driven exploration in discrete action-spaces. In this method, every action is associated with a degree of uncertainty, and the best action is selected based on a modified reward that accommodates the level of uncertainty. Similar work has been investigated within the the deep learning community, and proved to improve the overall performance[4].

  • [1] Osband, Ian, et al. "Deep exploration via bootstrapped DQN." Advances in neural information processing systems. 2016.
  • [2] Plappert, Matthias, et al. "Parameter space noise for exploration." arXiv preprint arXiv:1706.01905 (2017).
  • [3] Pathak, Deepak, et al. "Curiosity-driven exploration by self-supervised prediction." International Conference on Machine Learning (ICML). Vol. 2017. 2017.
  • [4] Kendall, Alex, and Yarin Gal. "What uncertainties do we need in bayesian deep learning for computer vision?." Advances in neural information processing systems. 2017.

Choosing the set of optimal hyperparameters is an important step in tuning a learning architecture. Existing methods inlude the naive grid-search or a random search over a given space of values. More statistical methods include Bayesian optimization [1,2],as well as gradient-based tuning [3]. However, with the recent evolved interest in reinforcement learning, it is now feasible to add a controller which is trained simultaneously with the proposed architecture, while the agent adaptively chooses new hyperparameter from a continuous action space[4]. In this thesis, you will work on formalizing the reinforcement learning task, by defining the state and reward associated with every action, (hyperparameter). At the first level, you will focus on simple models with a small number of hyperparameter, for example: regularization factor of an SVM, and then progress to adaptively tuning neural networks by selecting the right dropout values, L2-regularization factor, etc.

  • [1] Hutter, Frank, Holger H. Hoos, and Kevin Leyton-Brown. "Sequential model-based optimization for general algorithm configuration." International Conference on Learning and Intelligent Optimization. Springer, Berlin, Heidelberg, 2011.
  • [2] Snoek, Jasper, Hugo Larochelle, and Ryan P. Adams. "Practical bayesian optimization of machine learning algorithms." Advances in neural information processing systems. 2012.
  • [3] Bengio, Yoshua. "Gradient-based optimization of hyperparameters." Neural computation 12.8 (2000): 1889-1900.
  • [4] Lillicrap, Timothy P., et al. "Continuous control with deep reinforcement learning." arXiv preprint arXiv:1509.02971 (2015).

During the design of neural networks, many scientists and developers waste a huge amount of time and energy by testing several different kinds of configurations or by attempting to improve current existing models. In all of these situations, it often requires training the network from scratch, consuming a huge amount of time. To solve this problem, many methods attempt to use already existing models as a starting point and make small changes at a time. One of the current methods is presented by T. Chen et al. This method is known as Net2Net. By using knowledge transfer techniques based on function preserving transformations to replicate nodes or layers from a pre-trained-network, and training from that point, the network is guaranteed to either improve or maintain it's efficiency. This project aims to tackle this problem by formulating new methods of knowledge transfer. Your task is to understand the problem, formulate a new method while seeking to improve the state-of-the-art.

  • 1) Chen, Tianqi, Ian Goodfellow, and Jonathon Shlens. "Net2net: Accelerating learning via knowledge transfer." arXiv preprint arXiv:1511.05641 (2015).

In many machine learning problems, researchers often find themselves with the problem regarding incomplete or missing data. Several methods try to overcome this issue by using "complete" and previously annotated data while feeding models with only part of it. One of these scenarios is human motion classification where we are able to find databases with complete motion data information, but we might have only information from the arms (for example) in a hypothetical real-life application. To tackle this problem many researchers have used Deep Learning models to model, classify and even reconstruct the full body of the subject, however, only a few combine these tasks. In this project, your task is to understand the problem and the varied approaches and compare two of the best state-of-the-art techniques. In the end, an implementation and experiments should be presented. Also, a novel implementation is highly desirable.

  • 1) Catalin Ionescu, Dragos Papava, Vlad Olaru and Cristian Sminchisescu, Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 36, No. 7, July 2014
  • 2) Hanke, Sten, Matthieu Geist, and Andreas Holzinger. "Human Activity Recognition Using Recurrent Neural Networks." Machine Learning and Knowledge Extraction: First IFIP TC 5, WG 8.4, 8.9, 12.9 International Cross-Domain Conference, CD-MAKE 2017, Reggio, Italy, August 29–September 1, 2017, Proceedings. Vol. 10410. Springer, 2017.
  • 3) Ronao, Charissa Ann, and Sung-Bae Cho. "Human activity recognition with smartphone sensors using deep learning neural networks." Expert Systems with Applications 59 (2016): 235-244.

One of the biggest challenges in artificial intelligence is creating computer agents that are capable of making decisions as a human. A small sub-scenario of this area is simulating a human-player on a computer game. Several simple games have been already modeled, but complex and recent games are harder to model due to the high amount of information and features necessary to extract. Two different papers [1,2] implement two deep learning models that receive visual feedback and outputs actions as the buttons press. The models are trained based on human play sessions data, which includes the image on the screen and the buttons currently being pressed. Your job is to analyze different state-of-the-art papers, compare them and attempt to improve their current results. In the end, an implementation and experimental results should be presented. A novel improvement is highly desirable.

  • 1) Chen, Zhao, and Darvin Yi. "The Game Imitation: Deep Supervised Convolutional Networks for Quick Video Game AI." arXiv preprint arXiv:1702.05663 (2017).
  • 2) Chen, Zhao, and Darvin Yi. "The Game Imitation: Deep Supervised Convolutional Networks for Quick Video Game AI." arXiv preprint arXiv:1702.05663 (2017).

Recently several novel adaptive schemes for Coordinate Descent have emerged. For example [1] proposes to change the actual coordinate system over time such that it adapts to the loss surface. Another idea is to adaptively change the relative update frequencies of the coordinates [2,3], which is useful when the loss is very sensitive to some parameters, but very insensitive to others. Typically such adaptive methods yield superior performance, but suffer from scalability issues. The main objective in this thesis is to apply and specialize such adaptive optimization techniques to typical Machine Learning models as SVMs or DNNs. A secondary objective could be to create a Reinforcement-Learning Agent that learns how to select the best relative update frequencies.

  • Loshchilov, I., Schoenauer, M., & Sebag, M. (2011). Adaptive Coordinate Descent. In Proceedings of the 13th Annual Conference on Genetic and Evolutionary Computation (pp. 885–892). https://doi.org/10.1145/2001576.2001697
  • Glasmachers, T., & Dogan, U. (2013). Accelerated Coordinate Descent with Adaptive Coordinate Frequencies. In Asian Conference on Machine Learning (pp. 72–86). http://proceedings.mlr.press/v29/Glasmachers13.html
  • Perekrestenko, D., Cevher, V., & Jaggi, M. (2017). Faster Coordinate Descent via Adaptive Importance Sampling. In Artificial Intelligence and Statistics (pp. 869–877). http://proceedings.mlr.press/v54/perekrestenko17a.html

Game theoretical oriented approaches are becoming more and more popular in Machine Learning. In the n-player game, each agent (usually parametrized bz a neural network) tries to minimize its own objective function. The objective functions of a pair of agents can be opposing - e.g. agent A tries to minimize f, and agent B tries to minimize -f - (competitive case), they can be aligned - e.g. (cooperative case) or something in between. The goal of the experimenter is to find a Nash-equilibrium, i.e. a stable point in the sense that no agent can gain from changing its strategy. A prime example of a two player game are Generative Adversarial Nets (GANs) [3]. Here the two agents are the discriminator network, which tries to distinguish between real data and data generated by the second agent, the generator network whoose goal is to fool the discriminator. One of the biggest challenges in training GANs is the instability of the training procedure. It can be shown that the standard procedure of training GANs called Gradient Descent Ascent - simply alternatingly applying gradient updates to the generator/discriminator - does not necessarily converge. Recently some new ideas [1,2] have emerged to compute the Nash-equilibria in a more stable manner, for example Competitive Gradient Descent. [1] The goal of the master thesis is to understand and implement Competitive Gradient Descent [1], applying it to a novel setting and investigating possible improvements.

  • Schäfer, Florian, and Anima Anandkumar. “Competitive Gradient Descent.” ArXiv:1905.12103 [Cs, Math], May 28, 2019. http://arxiv.org/abs/1905.12103
  • Balduzzi, David, Sebastien Racaniere, James Martens, Jakob Foerster, Karl Tuyls, and Thore Graepel. “The Mechanics of N-Player Differentiable Games.” In International Conference on Machine Learning, 354–63, 2018. http://proceedings.mlr.press/v80/balduzzi18a.html
  • Goodfellow, Ian, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. “Generative Adversarial Nets.” In Advances in Neural Information Processing Systems 2014. http://papers.nips.cc/paper/5423-generative-adversarial-nets.pdf .
  • The inner objective: Given fixed hyperparameters h, find the optimal model weights w * (h), i.e. those which minimizes the loss on the training data
  • The outer objective: Find the hyperparameters h * for which the model with weights w * (h * ) has minimal loss on the validation data
  • Fu, Jie, Hongyin Luo, Jiashi Feng, Kian Hsiang Low, and Tat-Seng Chua “DrMAD: Distilling Reverse-Mode Automatic Differentiation for Optimizing Hyperparameters of Deep Neural Networks.” ArXiv:1601.00917 [Cs], January 5, 2016. http://arxiv.org/abs/1601.00917
  • Pedregosa, Fabian. “Hyperparameter Optimization with Approximate Gradient.” In International Conference on Machine Learning, 737–46, 2016. http://proceedings.mlr.press/v48/pedregosa16.html
  • Franceschi, Luca, Michele Donini, Paolo Frasconi, and Massimiliano Pontil. “Forward and Reverse Gradient-Based Hyperparameter Optimization.” ArXiv:1703.01785 [Stat], March 6, 2017. http://arxiv.org/abs/1703.01785
  • Franceschi, Luca, Paolo Frasconi, Saverio Salzo, Riccardo Grazzi, and Massimiliano Pontil. “Bilevel Programming for Hyperparameter Optimization and Meta-Learning.” In International Conference on Machine Learning, 1568–77, 2018. http://proceedings.mlr.press/v80/franceschi18a.html

Recent advances in Gradient based Hyperparameter Optimization allow the efficient optmization of large amounts of hyperparameters. Yet many questions are unexplored: How useful is it to optimize many hyperparameters? How to void overfitting on the validation data? What are interesting new model architectures to experiment with? In this thesis, you will perform exploratory work with gradient based hyperparameter optimization. A strong background in linear algebra and differential calculus is required.

  • Lorraine, Jonathan, Paul Vicol, and David Duvenaud. "Optimizing Millions of Hyperparameters by Implicit Differentiation." arXiv preprint arXiv:1911.02590 (2019)

Traditional recommender systems assume that the user profile and item attributes are static factors. This assumption doesn't consider the temporal dynamics that affect the user profile or the items e.g. user interest in an item may change based on the season of the year like watching Christmas movies in summer and winter. This thesis will identify the potential approaches to build a dynamic recommender system model that can predict users interest in items based on their historical interactions and behaviors.

  • 1) Chao-Yuan Wu, Amr Ahmed, Alex Beutel, Alexander J. Smola, and How Jing. 2017. Recurrent Recommender Networks. In Proceedings of the Tenth ACM International Conference on Web Search and Data Mining (WSDM '17). ACM, New York, NY, USA, 495-503. DOI: https://doi.org/10.1145/3018661.3018689
  • 2) Yongfeng Zhang, Min Zhang, Yi Zhang, Guokun Lai, Yiqun Liu, Honghui Zhang, and Shaoping Ma. 2015. Daily-Aware Personalized Recommendation based on Feature-Level Time Series Analysis. In Proceedings of the 24th International Conference on World Wide Web (WWW '15). International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, Switzerland, 1373-1383. DOI: https://doi.org/10.1145/2736277.2741087

The task of classifying multi-relational data spans a widerange of domains such as document classification in cita-tion networks, classification of emails and protein labeling inproteins interaction graphs. Current state-of-art classifica-tion models rely on learning per-entity latent representationsby mining the whole structure of the relations graph, how-ever, they still face two major problems. Firstly, it is verychallenging to generate expressive latent representations insparse multi-relational settings with implicit feedback rela-tions as there is very little information per-entity. Secondly,for entities with structured properties such as titles and ab-stracts (text) in documents, models have to be modified ad-hoc. This thesis will identify the potential deep learning approaches to predict entities labels based on their relations and interaction information such as predicting document's category in citation networks or prediciting user interest in social networks.

  • Cai, Hongyun, Vincent W. Zheng, and Kevin Chang. "A comprehensive survey of graph embedding: problems, techniques and applications." IEEE Transactions on Knowledge and Data Engineering (2018).
  • Kipf, Thomas N., and Max Welling. "Semi-supervised classification with graph convolutional networks." arXiv preprint arXiv:1609.02907 (2016).

Siamese networks have grown quite popular in the field of Machine Learning specially with research about object tracking and text equivalence, since they are able to compare samples and estimate their similarity. The hypothesis is that this architecture can be proven as a useful tool if you want to compare similarities between features from different data-sets becoming an asset for performing Model Agnostic Meta-Learning across tasks with different schemas.

In this thesis you will implement a siamese network to improve the performance of Model Agnostic Meta-Learning Methods that work across tasks with different schemas.

  • 1. Zhang, Zhipeng, and Houwen Peng. "Deeper and wider siamese networks for real-time visual tracking." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2019.
  • 2. Mueller, J., Thyagarajan, A. (2016, March). Siamese recurrent architectures for learning sentence similarity. In thirtieth AAAI conference on artificial intelligence.
  • 3. Brinkmeyer, L., Drumond, R. R., Scholz, R., Grabocka, J., Schmidt-Thieme, L. (2019). Chameleon: Learning model initializations across tasks with different schemas.
  • 4. Drumond, Rafael Rego, et al. "HIDRA: Head Initialization across Dynamic targets for Robust Architectures." Proceedings of the 2020 SIAM International Conference on Data Mining. Society for Industrial and Applied Mathematics, 2020.

The classical vehicle routing problem (VRP) designs optimal delivery routes where each vehicle only travels one route, each vehicle has the same characteristics and there is only one central depot. The goal of the VRP is to find a set of least-cost vehicle routes such that each customer is visited exactly once by one vehicle, each vehicle starts and ends its route at the depot, and the capacity of the vehicles is not exceeded. This classical VRP has been extended in many ways by introducing additional real-life aspects or characteristics, resulting in a large number of variants of the VRP. In the standard periodic vehicle routing problem (PVRP), customers require visits on one or more days within a planning period, and there are a set of feasible visit options for each customer. Customers must be assigned to a feasible visit option and a VRP is solved for each day in the planning period. The typical objective is to minimize the total distance traveled over the planning period. The PVRP arises in a diverse array of applications, from the collection of recyclables, to the routing of home healthcare nurses, to the collection of data in wireless networks. The wide applicability and versatility of the problem, coupled with the problem’s difficulty, has led to continuing interest and research efforts.

In computer science, evolutionary computation is a family of algorithms for global optimization inspired by biological evolution, and the subfield of artificial intelligence and soft computing studying these algorithms. In technical terms, they are a family of population-based trial and error problem solvers with a metaheuristic or stochastic optimization character. In evolutionary computation, an initial set of candidate solutions is generated and iteratively updated. Each new generation is produced by stochastically removing less desired solutions, and introducing small random changes. In biological terminology, a population of solutions is subjected to natural selection (or artificial selection) and mutation. As a result, the population will gradually evolve to increase in fitness, in this case the chosen fitness function of the algorithm. Evolutionary computation techniques can produce highly optimized solutions in a wide range of problem settings, making them popular in computer science.

  • 1. Braekers, Kris, Katrien Ramaekers, and Inneke Van Nieuwenhuyse. “The Vehicle Routing Problem: State of the Art Classification and Review.” Computers and Industrial Engineering 99 (September 1, 2016): 300–313. https://doi.org/10.1016/j.cie.2015.12.007.
  • 2. Campbell, Ann Melissa, and Jill Hardin Wilson. “Forty Years of Periodic Vehicle Routing.” Networks 63, no. 1 (January 2014): 2–15. https://doi.org/10.1002/net.21527.
  • 3. Zhou, Aimin, Bo-Yang Qu, Hui Li, Shi-Zheng Zhao, Ponnuthurai Nagaratnam Suganthan, and Qingfu Zhang. “Multiobjective Evolutionary Algorithms: A Survey of the State of the Art.” Swarm and Evolutionary Computation 1, no. 1 (March 1, 2011): 32–49. https://doi.org/10.1016/j.swevo.2011.03.001.
  • 4. Ombuki, Beatrice, Brian J. Ross, and Franklin Hanshar. “Multi-Objective Genetic Algorithms for Vehicle Routing Problem with Time Windows.” Applied Intelligence 24, no. 1 (February 1, 2006): 17–30. https://doi.org/10.1007/s10489-006-6926-z.

Reinforcement learning (RL) is an area of machine learning concerned with how software agents ought to take actions in an environment so as to maximize some notion of cumulative reward. In the operations research and control literature, reinforcement learning is called approximate dynamic programming, or neuro-dynamic programming (NDP). It describes a recent methodology that deals with the approximate solution of large and complex dynamic programming (DP) problems. Accordingly NDP combines simulation, learning, approximation architectures (e.g., neural networks) and the central ideas of DP to break the curse of dimensionality that is typical of DP. Therefore NDP could be a promising approach to solve large-scale and complex VRP.

  • 3. Sutton, Richard S, and Andrew G Barto. “Reinforcement Learning: An Introduction,”; https://web.stanford.edu/class/psych209/Readings/SuttonBartoIPRLBook2ndEd.pdf
  • 4. Secomandi, Nicola. “Comparing Neuro-Dynamic Programming Algorithms for the Vehicle Routing Problem with Stochastic Demands.” Computers and Operations Research 27, no. 11 (September 1, 2000): 1201–25. https://doi.org/10.1016/S0305-0548(99)00146-X.

In general reinforcement learning, action space is pre-defined (number of actions are fixed throughout the task). But, in real world situations, we may need to use new actions to accomplish the tasks. It can also be observed in the video games where the number of actions may increase with increase in level of the game.

As an example, consider a well known DAVE game. In level one, there are the actions {up}, {left}, {right}. Once, we enter level three, we will get an additional action of {shooting}. If we do not know that in future we may get new action, one has to learn new action, and may need to train the reinforcement learning algorithm from scratch with the additional action. This will take lot of time and the knowledge gained in previous levels goes waste.

In this thesis, a novel learning technique that can adapt the new action incrementally not from the scratch. The algorithm shall not lose the existing knowledge (ex: knowledge of previous levels in video games).

  • 1. Venkatesan, Rajasekar, and Meng Joo Er. "A novel progressive learning technique for multi-class classification." Neurocomputing 207 (2016): 310-321.
  • 2. Mnih, Volodymyr, et al. "Human-level control through deep reinforcement learning." Nature 518.7540 (2015): 529.

In general, while training a classification algorithm, we have true labels (Y) and predicted predicted labels (Y'). General trend of updating the parameters of the algorithm is by back propogating the loss which is average of absolute difference (or RMS loss). However, recent works shows that considereing top-k loss [1], average top-k loss [2] provide better results compared with traditional method.

In this work, one would analyse the results using above methods and find a better distribution over the losses for increasing the accuracy of the algorithm.

For information:

Top-k loss: It is the k^{th} maximum value in |Y - Y'| (similarly in RMS loss)

Average Top-k loss: It is the average of top-k values in |Y - Y'| (similarly in RMS loss)

  • 1. S. Shalev-Shwartz and Y. Wexler. Minimizing the maximal loss: How and why. In ICML, 2016
  • 2. Yanbo Fan, Siwei Lyu, Yiming Ying, Bao-Gang Hu. Learning with Average Top-k Loss. In NeruIPS, 2017
  • 3. Leonard Berrada, Andrew Zisserman, M. Pawan Kumar. Smooth Loss Functions for Deep Top-k Classification. in ICLR 2018

It is very difficult to automaticly extracting quantifying information out of medical images. Especially brain tumors can have unusal shapes so that common medical metrics do not work anymore (e.g. size of tumor). Their unpredictable appearance and shape make them challenging to be segmented in multi-modal brain imaging. Delineating brain tumor boundaries from magnetic resonance images is an essential task for the analysis of brain cancer. Several Machine Learning methods are used to automatically segment brain tumors.

If you are interested in other topics, please ask one of us directly.

machine learning Recently Published Documents

Total documents.

  • Latest Documents
  • Most Cited Documents
  • Contributed Authors
  • Related Sources
  • Related Keywords

An explainable machine learning model for identifying geographical origins of sea cucumber Apostichopus japonicus based on multi-element profile

A comparison of machine learning- and regression-based models for predicting ductility ratio of rc beam-column joints, alexa, is this a historical record.

Digital transformation in government has brought an increase in the scale, variety, and complexity of records and greater levels of disorganised data. Current practices for selecting records for transfer to The National Archives (TNA) were developed to deal with paper records and are struggling to deal with this shift. This article examines the background to the problem and outlines a project that TNA undertook to research the feasibility of using commercially available artificial intelligence tools to aid selection. The project AI for Selection evaluated a range of commercial solutions varying from off-the-shelf products to cloud-hosted machine learning platforms, as well as a benchmarking tool developed in-house. Suitability of tools depended on several factors, including requirements and skills of transferring bodies as well as the tools’ usability and configurability. This article also explores questions around trust and explainability of decisions made when using AI for sensitive tasks such as selection.

Automated Text Classification of Maintenance Data of Higher Education Buildings Using Text Mining and Machine Learning Techniques

Data-driven analysis and machine learning for energy prediction in distributed photovoltaic generation plants: a case study in queensland, australia, modeling nutrient removal by membrane bioreactor at a sewage treatment plant using machine learning models, big five personality prediction based in indonesian tweets using machine learning methods.

<span lang="EN-US">The popularity of social media has drawn the attention of researchers who have conducted cross-disciplinary studies examining the relationship between personality traits and behavior on social media. Most current work focuses on personality prediction analysis of English texts, but Indonesian has received scant attention. Therefore, this research aims to predict user’s personalities based on Indonesian text from social media using machine learning techniques. This paper evaluates several machine learning techniques, including <a name="_Hlk87278444"></a>naive Bayes (NB), K-nearest neighbors (KNN), and support vector machine (SVM), based on semantic features including emotion, sentiment, and publicly available Twitter profile. We predict the personality based on the big five personality model, the most appropriate model for predicting user personality in social media. We examine the relationships between the semantic features and the Big Five personality dimensions. The experimental results indicate that the Big Five personality exhibit distinct emotional, sentimental, and social characteristics and that SVM outperformed NB and KNN for Indonesian. In addition, we observe several terms in Indonesian that specifically refer to each personality type, each of which has distinct emotional, sentimental, and social features.</span>

Compressive strength of concrete with recycled aggregate; a machine learning-based evaluation

Temperature prediction of flat steel box girders of long-span bridges utilizing in situ environmental parameters and machine learning, computer-assisted cohort identification in practice.

The standard approach to expert-in-the-loop machine learning is active learning, where, repeatedly, an expert is asked to annotate one or more records and the machine finds a classifier that respects all annotations made until that point. We propose an alternative approach, IQRef , in which the expert iteratively designs a classifier and the machine helps him or her to determine how well it is performing and, importantly, when to stop, by reporting statistics on a fixed, hold-out sample of annotated records. We justify our approach based on prior work giving a theoretical model of how to re-use hold-out data. We compare the two approaches in the context of identifying a cohort of EHRs and examine their strengths and weaknesses through a case study arising from an optometric research problem. We conclude that both approaches are complementary, and we recommend that they both be employed in conjunction to address the problem of cohort identification in health research.

Export Citation Format

Share document.

Google Custom Search

Wir verwenden Google für unsere Suche. Mit Klick auf „Suche aktivieren“ aktivieren Sie das Suchfeld und akzeptieren die Nutzungsbedingungen.

Hinweise zum Einsatz der Google Suche

Technical University of Munich

  • Data Analytics and Machine Learning Group
  • TUM School of Computation, Information and Technology
  • Technical University of Munich

Technical University of Munich

Open Topics

We offer multiple Bachelor/Master theses, Guided Research projects and IDPs in the area of data mining/machine learning. A  non-exhaustive list of open topics is listed below.

If you are interested in a thesis or a guided research project, please send your CV and transcript of records to Prof. Stephan Günnemann via email and we will arrange a meeting to talk about the potential topics.

Generative Models for Drug Discovery

Type:  Mater Thesis / Guided Research

Prerequisites:

  • Strong machine learning knowledge
  • Proficiency with Python and deep learning frameworks (PyTorch or TensorFlow)
  • Knowledge of graph neural networks (e.g. GCN, MPNN)
  • No formal education in chemistry, physics or biology needed!

Description:

Effectively designing molecular geometries is essential to advancing pharmaceutical innovations, a domain which has experienced great attention through the success of generative models. These models promise a more efficient exploration of the vast chemical space and generation of novel compounds with specific properties by leveraging their learned representations, potentially leading to the discovery of molecules with unique properties that would otherwise go undiscovered. Our topics lie at the intersection of generative models like diffusion/flow matching models and graph representation learning, e.g., graph neural networks. The focus of our projects can be model development with an emphasis on downstream tasks ( e.g., diffusion guidance at inference time ) and a better understanding of the limitations of existing models.

Contact :  Johanna Sommer , Leon Hetzel

References:

Equivariant Diffusion for Molecule Generation in 3D

Equivariant Flow Matching with Hybrid Probability Transport for 3D Molecule Generation

Structure-based Drug Design with Equivariant Diffusion Models

Data Pruning and Active Learning

Type: Interdisciplinary Project (IDP) / Hiwi / Guided Research / Master's Thesis

  • Strong knowledge in machine learning
  • Very good coding skills
  • Proficiency with Python and deep learning frameworks (TensorFlow or PyTorch)

Data pruning and active learning are vital techniques in scaling machine learning applications efficiently. Data pruning involves the removal of redundant or irrelevant data, which enables training models with considerably less data but the same performance. Similarly, active learning describes the process of selecting the most informative data points for labeling, thus reducing annotation costs and accelerating model training. However, current methods are often computationally expensive, which makes them difficult to apply in practice. Our objective is to scale active learning and data pruning methods to large datasets using an extrapolation-based approach.

Contact: Sebastian Schmidt , Tom Wollschläger , Leo Schwinn

  • Large-scale Dataset Pruning with Dynamic Uncertainty

Efficient Machine Learning: Pruning, Quantization, Distillation, and More - DAML x Pruna AI

Type: Master's Thesis / Guided Research / Hiwi

The efficiency of machine learning algorithms is commonly evaluated by looking at target performance, speed and memory footprint metrics. Reduce the costs associated to these metrics is of primary importance for real-world applications with limited ressources (e.g. embedded systems, real-time predictions). In this project, you will work in collaboration with the DAML research group and the Pruna AI startup on investigating solutions to improve the efficiency of machine leanring models by looking at multiple techniques like pruning, quantization, distillation, and more.

Contact: Bertrand Charpentier

  • The Efficiency Misnomer
  • A Gradient Flow Framework for Analyzing Network Pruning
  • Distilling the Knowledge in a Neural Network
  • A Survey of Quantization Methods for Efficient Neural Network Inference

Deep Generative Models

Type:  Master Thesis / Guided Research

  • Strong machine learning and probability theory knowledge
  • Knowledge of generative models and their basics (e.g., Normalizing Flows, Diffusion Models, VAE)
  • Optional: Neural ODEs/SDEs, Optimal Transport, Measure Theory

With recent advances, such as Diffusion Models, Transformers, Normalizing Flows, Flow Matching, etc., the field of generative models has gained significant attention in the machine learning and artificial intelligence research community. However, many problems and questions remain open, and the application to complex data domains such as graphs, time series, point processes, and sets is often non-trivial. We are interested in supervising motivated students to explore and extend the capabilities of state-of-the-art generative models for various data domains.

Contact : Marcel Kollovieh , Marten Lienen ,  David Lüdke

  • Flow Matching for Generative Modeling
  • Auto-Encoding Variational Bayes
  • Denoising Diffusion Probabilistic Models 
  • Structured Denoising Diffusion Models in Discrete State-Spaces

Graph Structure Learning

Type:  Guided Research / Hiwi

  • Optional: Knowledge of graph theory and mathematical optimization

Graph deep learning is a powerful ML concept that enables the generalisation of successful deep neural architectures to non-Euclidean structured data. Such methods have shown promising results in a vast range of applications spanning the social sciences, biomedicine, particle physics, computer vision, graphics and chemistry. One of the major limitations of most current graph neural network architectures is that they often rely on the assumption that the underlying graph is known and fixed. However, this assumption is not always true, as the graph may be noisy or partially and even completely unknown. In the case of noisy or partially available graphs, it would be useful to jointly learn an optimised graph structure and the corresponding graph representations for the downstream task. On the other hand, when the graph is completely absent, it would be useful to infer it directly from the data. This is particularly interesting in inductive settings where some of the nodes were not present at training time. Furthermore, learning a graph can become an end in itself, as the inferred structure can provide complementary insights with respect to the downstream task. In this project, we aim to investigate solutions and devise new methods to construct an optimal graph structure based on the available (unstructured) data.

Contact : Filippo Guerranti

  • A Survey on Graph Structure Learning: Progress and Opportunities
  • Differentiable Graph Module (DGM) for Graph Convolutional Networks
  • Learning Discrete Structures for Graph Neural Networks

NodeFormer: A Scalable Graph Structure Learning Transformer for Node Classification

Graph Neural Networks

Type:  Master's thesis / Bachelor's thesis / guided research

  • Knowledge of graph/network theory

Graph neural networks (GNNs) have recently achieved great successes in a wide variety of applications, such as chemistry, reinforcement learning, knowledge graphs, traffic networks, or computer vision. These models leverage graph data by updating node representations based on messages passed between nodes connected by edges, or by transforming node representation using spectral graph properties. These approaches are very effective, but many theoretical aspects of these models remain unclear and there are many possible extensions to improve GNNs and go beyond the nodes' direct neighbors and simple message aggregation.

Contact: Simon Geisler

  • Semi-supervised classification with graph convolutional networks
  • Relational inductive biases, deep learning, and graph networks
  • Diffusion Improves Graph Learning
  • Weisfeiler and leman go neural: Higher-order graph neural networks
  • Reliable Graph Neural Networks via Robust Aggregation

Physics-aware Graph Neural Networks

Type:  Master's thesis / guided research

  • Proficiency with Python and deep learning frameworks (JAX or PyTorch)
  • Knowledge of graph neural networks (e.g. GCN, MPNN, SchNet)
  • Optional: Knowledge of machine learning on molecules and quantum chemistry

Deep learning models, especially graph neural networks (GNNs), have recently achieved great successes in predicting quantum mechanical properties of molecules. There is a vast amount of applications for these models, such as finding the best method of chemical synthesis or selecting candidates for drugs, construction materials, batteries, or solar cells. However, GNNs have only been proposed in recent years and there remain many open questions about how to best represent and leverage quantum mechanical properties and methods.

Contact: Nicholas Gao

  • Directional Message Passing for Molecular Graphs
  • Neural message passing for quantum chemistry
  • Learning to Simulate Complex Physics with Graph Network
  • Ab initio solution of the many-electron Schrödinger equation with deep neural networks
  • Ab-Initio Potential Energy Surfaces by Pairing GNNs with Neural Wave Functions
  • Tensor field networks: Rotation- and translation-equivariant neural networks for 3D point clouds

Robustness Verification for Deep Classifiers

Type: Master's thesis / Guided research

  • Strong machine learning knowledge (at least equivalent to IN2064 plus an advanced course on deep learning)
  • Strong background in mathematical optimization (preferably combined with Machine Learning setting)
  • Proficiency with python and deep learning frameworks (Pytorch or Tensorflow)
  • (Preferred) Knowledge of training techniques to obtain classifiers that are robust against small perturbations in data

Description : Recent work shows that deep classifiers suffer under presence of adversarial examples: misclassified points that are very close to the training samples or even visually indistinguishable from them. This undesired behaviour constraints possibilities of deployment in safety critical scenarios for promising classification methods based on neural nets. Therefore, new training methods should be proposed that promote (or preferably ensure) robust behaviour of the classifier around training samples.

Contact: Aleksei Kuvshinov

References (Background):

  • Intriguing properties of neural networks
  • Explaining and harnessing adversarial examples
  • SoK: Certified Robustness for Deep Neural Networks
  • Certified Adversarial Robustness via Randomized Smoothing
  • Formal guarantees on the robustness of a classifier against adversarial manipulation
  • Towards deep learning models resistant to adversarial attacks
  • Provable defenses against adversarial examples via the convex outer adversarial polytope
  • Certified defenses against adversarial examples
  • Lipschitz-margin training: Scalable certification of perturbation invariance for deep neural networks

Uncertainty Estimation in Deep Learning

Type: Master's Thesis / Guided Research

  • Strong knowledge in probability theory

Safe prediction is a key feature in many intelligent systems. Classically, Machine Learning models compute output predictions regardless of the underlying uncertainty of the encountered situations. In contrast, aleatoric and epistemic uncertainty bring knowledge about undecidable and uncommon situations. The uncertainty view can be a substantial help to detect and explain unsafe predictions, and therefore make ML systems more robust. The goal of this project is to improve the uncertainty estimation in ML models in various types of task.

Contact: Tom Wollschläger ,   Dominik Fuchsgruber ,   Bertrand Charpentier

  • Can You Trust Your Model’s Uncertainty? Evaluating Predictive Uncertainty Under Dataset Shift
  • Predictive Uncertainty Estimation via Prior Networks
  • Posterior Network: Uncertainty Estimation without OOD samples via Density-based Pseudo-Counts
  • Evidential Deep Learning to Quantify Classification Uncertainty
  • Weight Uncertainty in Neural Networks

Hierarchies in Deep Learning

Type:  Master's Thesis / Guided Research

Multi-scale structures are ubiquitous in real life datasets. As an example, phylogenetic nomenclature naturally reveals a hierarchical classification of species based on their historical evolutions. Learning multi-scale structures can help to exhibit natural and meaningful organizations in the data and also to obtain compact data representation. The goal of this project is to leverage multi-scale structures to improve speed, performances and understanding of Deep Learning models.

Contact: Marcel Kollovieh , Bertrand Charpentier

  • Tree Sampling Divergence: An Information-Theoretic Metricfor Hierarchical Graph Clustering
  • Hierarchical Graph Representation Learning with Differentiable Pooling
  • Gradient-based Hierarchical Clustering
  • Gradient-based Hierarchical Clustering using Continuous Representations of Trees in Hyperbolic Space

M.Tech/Ph.D Thesis Help in Chandigarh | Thesis Guidance in Chandigarh

machine learning thesis topics

[email protected]

machine learning thesis topics

+91-9465330425

machine learning thesis topics

Latest thesis topics in Machine Learning for research scholars:

Choosing a research and thesis topics in Machine Learning is the first choice of masters and Doctorate scholars now a days. Though, choosing and working on a thesis topic in machine learning is not an easy task as Machine learning uses certain statistical algorithms to make computers work in a certain way without being explicitly programmed. The algorithms receive an input value and predict an output for this by the use of certain statistical methods. The main aim of machine learning is to create intelligent machines which can think and work like human beings. Achieving the above mentioned goals is surely not very easy because of which students who choose research topic in machine learning face difficult challenges and require professional thesis help in their thesis work.

Below is the list of the latest thesis topics in Machine learning for research scholars:

  • The classification technique for the face spoof detection in artificial neural networks using concepts of machine learning .
  • The iris detection and reorganization system using classification and glcm algorithm in machine learning.
  • Using machine learning algorithms in the detection of pattern system using algorithm of textual feature analysis and classification
  • The plant disease detection using glcm and KNN classification in neural networks merged with the concepts of machine learning
  • Using the algorithms of machine learning to propose technique for the prediction analysis in data mining
  • The sentiment analysis technique using SVM classifier in data mining using machine learning approach
  • The heart disease prediction using technique of classification in machine learning using the concepts of data mining.

So let’s start with machine learning.

First of all…

What exactly is machine learning?

Find the link at the end to download the latest topics for thesis and research in Machine Learning

What is Machine Learning?

machine learning thesis topics

Machine Learning is a branch of artificial intelligence that gives systems the ability to learn automatically and improve themselves from the experience without being explicitly programmed or without the intervention of human. Its main aim is to make computers learn automatically from the experience.

Requirements of creating good machine learning systems

So what is required for creating such machine learning systems? Following are the things required in creating such machine learning systems:

Data – Input data is required for predicting the output.

Algorithms – Machine Learning is dependent on certain statistical algorithms to determine data patterns.

Automation – It is the ability to make systems operate automatically.

Iteration – The complete process is iterative i.e. repetition of process.

Scalability – The capacity of the machine can be increased or decreased in size and scale.

Modeling – The models are created according to the demand by the process of modeling.

Methods of Machine Learning

machine learning thesis topics

Machine Learning methods are classified into certain categories These are:

  • Supervised Learning
  • Unsupervised Learning

Reinforcement Learning

Supervised Learning – In this method, input and output is provided to the computer along with feedback during the training. The accuracy of predictions by the computer during training is also analyzed. The main goal of this training is to make computers learn how to map input to the output.

Unsupervised Learning – In this case, no such training is provided leaving computers to find the output on its own. Unsupervised learning is mostly applied on transactional data. It is used in more complex tasks. It uses another approach of iteration known as deep learning to arrive at some conclusions.

Reinforcement Learning – This type of learning uses three components namely – agent, environment, action. An agent is the one that perceives its surroundings, an environment is the one with which an agent interacts and acts in that environment. The main goal in reinforcement learning is to find the best possible policy.

How does machine learning work?

machine learning thesis topics

Machine learning makes use of processes similar to that of data mining. Machine learning algorithms are described in terms of target function(f) that maps input variable (x) to an output variable (y). This can be represented as:

There is also an error e which is the independent of the input variable x. Thus the more generalized form of the equation is:

In machine the mapping from x to y is done for predictions. This method is known as predictive modeling to make most accurate predictions. There are various assumptions for this function.

Benefits of Machine Learning

mtech thesis topics in machine learning

Everything is dependent on machine learning. Find out what are the benefits of machine learning.

Decision making is faster – Machine learning provides the best possible outcomes by prioritizing the routine decision-making processes.

Adaptability – Machine Learning provides the ability to adapt to new changing environment rapidly. The environment changes rapidly due to the fact that data is being constantly updated.

Innovation – Machine learning uses advanced algorithms that improve the overall decision-making capacity. This helps in developing innovative business services and models.

Insight – Machine learning helps in understanding unique data patterns and based on which specific actions can be taken.

Business growth – With machine learning overall business process and workflow will be faster and hence this would contribute to the overall business growth and acceleration.

Outcome will be good – With machine learning the quality of the outcome will be improved with lesser chances of error.

Branches of Machine Learning

  • Computational Learning Theory
  • Adversarial Machine Learning
  • Quantum Machine Learning
  • Robot Learning
  • Meta-Learning

Computational Learning Theory – Computational learning theory is a subfield of machine learning for studying and analyzing the algorithms of machine learning. It is more or less similar to supervised learning.

Adversarial Machine Learning – Adversarial machine learning deals with the interaction of machine learning and computer security. The main aim of this technique is to look for safer methods in machine learning to prevent any form of spam and malware. It works on the following three principles:

Finding vulnerabilities in machine learning algorithms.

Devising strategies to check these potential vulnerabilities.

Implementing these preventive measures to improve the security of the algorithms.

Quantum Machine Learning – This area of machine learning deals with quantum physics. In this algorithm, the classical data set is translated into quantum computer for quantum information processing. It uses Grover’s search algorithm to solve unstructured search problems.

Predictive Analysis – Predictive Analysis uses statistical techniques from data modeling, machine learning and data mining to analyze current and historical data to predict the future. It extracts information from the given data. Customer relationship management(CRM) is the common application of predictive analysis.

Robot Learning – This area deals with the interaction of machine learning and robotics. It employs certain techniques to make robots to adapt to the surrounding environment through learning algorithms.

Grammar Induction – It is a process in machine learning to learn formal grammar from a given set of observations to identify characteristics of the observed model. Grammar induction can be done through genetic algorithms and greedy algorithms.

Meta-Learning – In this process learning algorithms are applied on meta-data and mainly deals with automatic learning algorithms.

Best Machine Learning Tools

Here is a list of artificial intelligence and machine learning tools for developers:

ai-one – It is a very good tool that provides software development kit for developers to implement artificial intelligence in an application.

Protege – It is a free and open-source framework and editor to build intelligent systems with the concept of ontology. It enables developers to create, upload and share applications.

IBM Watson – It is an open-API question answering system that answers questions asked in natural language. It has a collection of tools which can be used by developers and in business.

DiffBlue – It is another tool in artificial intelligence whose main objective is to locate bugs, errors and fix weaknesses in the code. All such things are done through automation.

TensorFlow – It is an open-source software library for machine learning. TensorFlow provides a library of numerical computations along with documentation, tutorials and other resources for support.

Amazon Web Services – Amazon has launched toolkits for developers along with applications which range from image interpretation to facial recognition.

OpenNN – It is an open-source, high-performance library for advanced analytics and is written in C++ programming language. It implements neural networks. It has a lot of tutorials and documentation along with an advanced tool known as Neural Designer.

Apache Spark – It is a framework for large-scale processing of data. It also provides a programming tool for deep learning on various machines.

Caffe – It is a framework for deep learning and is used in various industrial applications in the area of speech, vision and expression.

Veles – It is another deep learning platform written in C++ language and make use of python language for interaction between the nodes.

Machine Learning Applications

Following are some of the applications of machine learning:

Cognitive Services

Medical Services

Language Processing

Business Management

Image Recognition

Face Detection

Video Games

Computer Vision

Pattern Recognition

Machine Learning in Bioinformatics

Bioinformatics term is a combination of two terms bio, informatics. Bio means related to biology and informatics means information. Thus bioinformatics is a field that deals with processing and understanding of biological data using computational and statistical approach. Machine Learning has a number of applications in the area of bioinformatics. Machine Learning find its application in the following subfields of bioinformatics:

Genomics – Genomics is the study of DNA of organisms. Machine Learning systems can help in finding the location of protein-encoding genes in a DNA structure. Gene prediction is performed by using two types of searches named as extrinsic and intrinsic. Machine Learning is used in problems related to DNA alignment.

Proteomics – Proteomics is the study of proteins and amino acids. Proteomics is applied to problems related to proteins like protein side-chain prediction, protein modeling, and protein map prediction.

Microarrays – Microarrays are used to collect data about large biological materials. Machine learning can help in the data analysis, pattern prediction and genetic induction. It can also help in finding different types of cancer in genes.

System Biology – It deals with the interaction of biological components in the system. These components can be DNA, RNA, proteins and metabolites. Machine Learning help in modeling these interactions.

Text mining – Machine learning help in extraction of knowledge through natural language processing techniques.

Deep Learning

machine learning thesis topics

Deep Learning is a part of the broader field machine learning and is based on data representation learning. It is based on the interpretation of artificial neural network. Deep Learning algorithm uses many layers of processing. Each layer uses the output of previous layer as an input to itself. The algorithm used can be supervised algorithm or unsupervised algorithm. Deep Learning is mainly developed to handle complex mappings of input and output. It is another hot topic for M.Tech thesis and project along with machine learning.

Deep Neural Network

Deep Neural Network is a type of Artificial Neural Network with multiple layers which are hidden between the input layer and the output layer. This concept is known as feature hierarchy and it tends to increase the complexity and abstraction of data. This gives network the ability to handle very large, high-dimensional data sets having millions of parameters. The procedure of deep neural networks is as follows:

Consider some examples from a sample dataset.

Calculate error for this network.

Improve weight of the network to reduce the error.

Repeat the procedure.

Applications of Deep Learning

Here are some of the applications of Deep Learning:

Automatic Speech Recognition

Natural Language Processing

Customer Relationship Management

Bioinformatics

Mobile Advertising

Advantages of Deep Learning

Deep Learning helps in solving certain complex problems with high speed which were earlier left unsolved. Deep Learning is very useful in real world applications. Following are some of the main advantages of deep learning:

Eliminates unnecessary costs – Deep Learning helps to eliminate unnecessary costs by detecting defects and errors in the system.

Identifies defects which otherwise are difficult to detect – Deep Learning helps in identifying defects which left untraceable in the system.

Can inspect irregular shapes and patterns – Deep Learning can inspect irregular shapes and patterns which is difficult for machine learning to detect.

From this introduction, you must have known that why this topic is called as hot for your M.Tech thesis and projects. This was just the basic introduction to machine learning and deep learning. There is more to explore in these fields. You will get to know more once you start doing research on this topic for your M.Tech thesis. You can get thesis assistance and guidance on this topic from experts specialized in this field.

Research and Thesis Topics in Machine Learning

Here is the list of current research and thesis topics in Machine Learning :

Machine Learning Algorithms

Supervised Machine Learning

Unsupervised Machine Learning

Neural Networks

Predictive Learning

Bayesian Network

Data Mining

For starting with Machine Learning, you need to know some algorithms. Machine Learning algorithms are classified into three categories which provide the base for machine learning. These categories of algorithms are supervised learning, unsupervised learning, and reinforcement learning. The choice of algorithms depends upon the type of tasks you want to be done along with the type, quality, and nature of data present. The role of input data is crucial in machine learning algorithms.

Computer Vision is a field that deals with making systems that can read and interpret images. In simple terms, computer vision is a method of transmitting human intelligence and vision in machines. In computer vision, data is collected from images which are imparted to systems. The system will take action according to the information it interprets from what it sees.

It is a good topic for machine learning masters thesis. It is a type of machine learning algorithm in which makes predictions based on known data-sets. Input and output is provided to the system along with feedback. Supervised Learning is further classified into classification and regression problems. In the classification problem, the output is a category while in regression problem the output is a real value.

It is another category of machine learning algorithm in which input is known but the output is not known. Prior training is not provided to the system as in case of supervised learning. The main purpose of unsupervised learning is to model the underlying structure of data. Clustering and Association are the two types of unsupervised learning problems. k-means and Apriori algorithm are the examples of unsupervised learning algorithms.

Deep Learning is a hot topic in Machine Learning. It is already explained above. It is a part of the family of machine learning and deals with the functioning of the artificial neural network. Neural Networks are used to study the functioning of the human brain. It is one of the growing and exciting field. Deep learning has made it possible for the practical implementation of various machine learning applications.

Neural Networks are the systems to study the biological neural networks. It is an important application of machine learning and a good topic for masters thesis and research. The main purpose of Artificial Neural Network is to study how the human brain works. It finds its application in computer vision, speech recognition, machine translation etc. Artificial Neural Network is a collection of nodes which represent neurons.

Reinforcement Learning is a category of machine learning algorithms. Reinforcement Learning deals with software agents to study how these agents take actions in an environment in order to maximize their performance. Reinforcement Learning is different from supervised learning in the sense that correct input and output parameters are not provided.

Predictive Learning is another good topic for thesis in machine learning. In this technique, a model is built by an agent of its environment in which it performs actions. There is another field known as predictive analytics which is used to make predictions about future events which are unknown. For this, techniques like data mining, statistics, modeling, machine learning, and artificial intelligence are used.

It is a network that represents probabilistic relationships via Directed Acyclic Graph(DAG). There are algorithms in Bayesian Network for inference and learning. In the network, a probability function is there for each node which takes an input to give probability to the value associated with the node. Bayesian Network finds its application in bioinformatics, image processing, and computational biology.

Data Mining is the process of finding patterns from large data-sets to extract valuable information to make better decisions. It is a hot area of research. This technology use method from machine learning, statistics, and database systems for processing. There exist data mining techniques like clustering, association, decision trees, classification for the data mining process.

Click on the following link to download the latest thesis and research topics in Machine Learning

Latest Thesis and Research Topics on Machine Learning(pdf)

For more details Contact Us.  You can call us on this number +91-9465330425 or drop an email at   [email protected]   for any type of dissertation help in India. You can also fill the query form on the website. 

You can also visit our website Techsparks and follow us on Pinterest , Facebook , Twitter, YouTube and Linkedin for latest thesis blog.

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Save my name, email, and website in this browser for the next time I comment.

Quick Enquiry

Get a quote, share your details to get free.

Kindson The Genius

Kindson The Genius

Providing the best learning experience for professionals

10 Machine Learning Project (Thesis) Topics for 2020

kindsonthegenius

Are you looking for some interesting project ideas for your thesis, project or dissertation? Then be sure that a machine learning topic would be a very good topic to write on. I have outlined 10 different topics. These topics are really good because you can easily obtain the dataset (i will provide the link to the dataset) and you can as well get some support from me. Let me know if you need any support in preparing your thesis.

You can leave a comment below in the comment area.

machine learning thesis topics

1.  Machine Learning Model for Classification and Detection of Breast Cancer (Classification)

The data is provided by the Oncology department and details instances and related attributes which are nine in all.

You can obtain the dataset from here

2. Intelligent Internet Ads Generation (Classification)

This is one of the most interesting topics for me. The reason is because the revenue generated or expended by ads campaign depends not just on the volume of the ads, but also on the relevance of the ads. Therefore it is possible to increase revenue and reduce spending by developing a Machine Learning model that select relevants ads with a high level of accuracy.  The dataset provides a collection of ads as well as the structure and geometry of the ads.

Get the ads dataset from here

3. Feature Extraction for National Census Data (Clustering)

This looks like big data stuff. But no! It’s simply dataset you can use for analysis. It is the actual data obtained from the US census in 1990. There are 68 attributes for each of the records and clustering would be performed to identify trends in the data.

You can obtain census the dataset from here

4. Movie Outcome Prediction (Classification)

This is quite a tasking project but its quite interesting. Before now, there exists models to predict the ratings of movies on a scale of 0 to 10 or 1 to 5. But this takes it a step further. You actually need to determine the outcome of the movie.  The data set is a large multivariate dataset of movie director, cast, individual roles of the actor, remarks, studio and relevant documents.

You can get the movies dataset from here

5. Forest Fire Area Coverage Prediction (Regression)

This project have been classified as difficult but I don’t think so. The objective to predict the the area affected by forest fires. Dataset include relevant meteological information and other parameters taken from a region of Portugal.

You can get the fire dataset from here

6. Atmospheric Ozone Level Analysis and Detection (Clustering)

Two ground ozone datasets are provided for this. Data includes temperatures at various times of the day as well as wind speed. The data included in the dataset was collected in a span of 6 years from 1998 to 2004.

You can get the Ozone dataset from here

7. Crime Prediction in New York City (Regression)

If you have watched the movie, ‘Person of Interest’ directed by Jonathan Nolan, then you will appreciate the fact that there is a possibility of predicting  violent criminal activities before they actually occur. Dataset would contain historical data on crime rate, types of crimes occurrence per region.

You can get the crime dataset from here

8. Sentiment Analysis on Amazon ECommerce User Reviews (Classification)

The dataset for this project is derived from user review comments from Amazon users. The model should be able to perform analysis on the training dataset and come up with a model that classifies the reviews based on sentiments. Granularity can be improved by generating predictions based on location and other factors.

You can get the reviews dataset from here

9. Home Eletrical Power Consumption Analysis (Regression)

Everyone uses electricity at home. Or rather, almost everyone! Would is not be great to have a system that helps to predict electricity consumption. Training dataset provided for this project includes feature set such as the size of the home, duration and more

You can get the dataset from here

10. Predictive Modelling of Individual Human Knowledge (Classification and Clustering)

Here the available dataset provide a collection of data about an individual on a subject matter. You are required to create a model that would try to quantify the amount of knowledge the individual have on the given subject. You can be creating by trying to also infer the performance of the user on certain exams.

I hope these 10 Machine Learning Project topic would be helpful to you.

Thanks for reading and do leave a comment below if you need some support

User Avatar

kindsonthegenius

Kindson Munonye is currently completing his doctoral program in Software Engineering in Budapest University of Technology and Economics

You might also like

Pca tutorial 2 – how to perform principal components analysis (pca), introduction to higher order singular value decomposition (hosvd), from programmer to data scientist – 5 steps, 2 thoughts on “ 10 machine learning project (thesis) topics for 2020 ”.

Is there any suggestion related to educational data mining?

I’m working on this. You can subscribe to my channel so when I make the update, you can get notified https://www.youtube.com/channel/UCvHgEAcw6VpcOA3864pSr5A

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Save my name, email, and website in this browser for the next time I comment.

PHD PRIME

Thesis Topics for Machine Learning

  Machine learning is one of the recently growing fields of research for classification, clustering, and prediction of input data . The techniques of ensembles and hybridization have contributed to the improvement of machine learning models . As a result, the speed of computation, operation, accuracy, and robustness of the machine learning models are enhanced. Through this article, you can get an overview of novel machine learning models, their design, performance, merits, and uses explained via a new taxonomic approach . Also, you can get all essential details regarding any thesis topics for machine learning from this page.

At present, there are many new ensembles and hybridized machine learning models being introduced and developed .  Here the essentials of thesis writing are presented to you by our world-class certified writers and developers. What are the essential elements of a thesis statement?

  • First of all you have to understand that thesis statement writing is the most crucial process which involves a lot of time and thinking
  • Enough research data and evidence have to be gathered before writing a thesis statement
  • The main Idea or the objective has to be presented clearly with supporting evidence
  • Also remember that the thesis statement should be in accordance with the argument where adjustments are allowed

Usually, research scholars interact with our writers and experts for all aspects of thesis writing in machine learning. So we insist that you contact us much before you start your thesis so that you can have a clear-cut vision and well-strategized approach towards writing the best thesis.

Top 5 Research Thesis Topics for Machine Learning

Let us now have an idea about various headings to be included in any thesis topics for machine learning.

  • Introduction – overview of the thesis
  • Related / Existing works – presents of existing research
  • Problems definition/statements – identify and highlight the problems
  • Research methodology – convey the proposed concepts
  • Results and Discussion – discuss the results of the proposed works with previous works
  • Conclusion and future work – present the results of the proposed work

The introduction is the very first part of your thesis. It is the way by which you tend to create the first impression in the minds of the readers. What are the parts of the introduction in the thesis?

  • The issue under examination is the core of an overview
  • Main Idea and assertion has to be mentioned clearly
  • Thesis statement and argument forms the fundamental aspect here
  • Address the audience to prove to them that they are at the right place
  • Scope of your paper should be mentioned satisfactorily
  • The Planning based approach that you used to conduct research

In general, the choice of words, tone, approach, and language decide the quality of a thesis likewise the introduction. Our technical team and expert writers have gained enough experience writing thesis topics in machine learning. The amount of field knowledge and expertise that we gathered is quite large which will be of great use to you. Let us now talk about the next most important topic of a thesis called the issue

What are the guidelines for thesis writing? 

Under the heading of the issue, the following aspects of research are to be included

  • The background history about your result issue or concern solving which is stated as your objective
  • The impact of the issue in this field
  • Important characteristic features that affect the issue
  • Potential research solutions that are undertaken for research

With the massive amount of reliable and authentic research materials that we provide, you can surely get all the necessary information to include in the issues part of your thesis. Also, our engineers and technical team are here to solve any kind of technical queries that you may get. Let us now talk about the literature review  

LITERATURE REVIEW 

  • With important references and constructs from standard textbooks journals and relevant publications you need to make the following descriptions
  • Relevant theory
  • Issue explanation
  • Potential solution
  • Theoretical constructs
  • Explanation on major theories
  • Empirical literature from journal articles are considered for the following aspects
  • Explanation on latest empirical studies
  • Summary of the methodology adopted
  • Important findings of the study
  • Constraints associated with your findings
  • The pathway of your research study has to be organized in line with the literature review to make keynotes on the following
  • The referred definitions and concepts
  • Unique aspects of the issues under examination
  • Suitable method of your research

If you are searching for the best and most reliable online research guide for all kinds of thesis topics in machine learning then you are here at the right place. You can get professional and customized research support aligned with your institutional format from our experts. Let us now look into the method section in detail below  

The following are the different aspects that you need to incorporate in the methods section of your thesis

  • The research questions and issues under your examination
  • Description of proposed works like data collection
  • Rationale justification for the method of your choice

In addition to these aspects, you need to provide a clear description of all the research methods that you adopt in your study. For this purpose here are our research experts who will provide you with details on novel and innovative approaches useful for your research . You can also get concise and precise quantitative research data from us. Let us now look into this section of results

RESULTS AND DISCUSSION

On the page of results and discussion you need to incorporate the following aspects

  • Description of major findings
  • Visualization tools like charts, graphs, and tables to present the findings
  • Relevant previous studies and results
  • Creative and new results that you obtained
  • Scopes to expand the previous studies with your findings
  • Constraints of your study

The support of technical experts can help you do the best research work in machine learning . The interested researcher plus reliable and experienced research support makes the best PhD work possible. With our guidance, you get access to the best combo needed to carry out your research. Let’s now discuss the conclusions part  

Conclusion and recommendation

In the part of conclusion, you need to include the following aspects

  • Recap of issues being discussed
  • Methods used and major findings
  • Comparison between the original objective and accomplished results
  • Scope for future expansion of your research

For each and every aspect of your machine learning PhD thesis , you can get complete support from our experts. In this respect let us now look to the topmost machine learning thesis topics below  

Top 5 Thesis Topics for machine learning

  • Machine learning is of great importance to physicians in the following perspectives
  • Chatbots for speech recognition
  • Pattern recognition for disease detection
  • Treatment recommendation
  • Detecting cancerous cells
  • Body fluid analysis
  • Identification of phenotypes in case of rare diseases
  • Classifying data into groups for fault detection is possible using machine learning
  • The following are some real-time examples for predictive analysis
  • Fraudulent and legitimate transaction
  • Improvement of prediction mechanism for detecting faults
  • From the basics of developing products to predicting the stock market and real estate prices, predictive analytics is of greater importance
  • Using a trading algorithm that makes use of a proper strategy for financing huge volumes of security is called statistical arbitrage
  • Real-time examples of statistical arbitrage
  • Analysis of huge data sets
  • Algorithm-based trading for market microstructural analysis
  • Real-time arbitrage possibilities
  • Machine learning is used to enhance the strategy for statistical arbitrage as a result of which advanced results can be obtained
  • In order to help the predictive analytics mechanisms to obtain increased accuracy feature extraction using machine learning plays a significant role
  • Dataset annotations can be performed with greater significance using machine learning extraction methods where structured data can be extracted from unstructured information
  • Real-time examples of machine learning-based feature extraction include the following
  • Vocal cord disorder prediction
  • Mechanism for prevention diagnosis and treatment of many disorders
  • Detecting and solving many physiological problems in a Swift manner
  • Extraction of critical information becomes easy with machine learning even when large volumes of data are being processed
  • Machine learning methodologies can be used for translating speech into texts
  • Recorded speech and real-time voice can be converted into text using machine learning systems designed for this purpose
  • Speech can also be classified based on intensity, time, and frequency
  • Voice search, appliance control, and voice dialing are the main real-time examples of speech recognition

In order to get confidential research guidance from world-class experts on all these thesis topics for machine learning, you can feel free to contact us. With more than 15 years of customer satisfaction, we are providing in-depth Research and advanced project support for all thesis topics for machine learning . Our thesis writing support also includes the following aspects

  • Multiple revisions
  • Complete grammatical check
  • Formatting and editing
  • Benchmark reference and citations from topmost journals
  • Work privacy
  • Internal review

We ensure all these criteria are conferred to you by world-class certified engineers, developers, and writers. So you can avail of our services with elevated confidence. We are here to support you fully. Let us now see some important machine learning methods in the following  

Machine learning methods

Machine learning techniques are most often used in cases of making automatic decisions for any kind of input that they are trained and implemented for. Therefore machine learning approaches are expected to support the following aspects in decision making.

  • Maximum accuracy of recommendations
  • In-depth understanding and analysis before deciding to increase the trustworthiness

The decision-making approach using machine learning methods provides for higher accuracy in prediction and advanced comprehensible models respectively in implicit and explicit learning. For all your doubts and queries regarding the above-mentioned machine learning and decision-making approaches , you may feel free to contact us at any time of your convenience. Our technical team is highly experienced and skilled in resolving any kind of queries . Let us now see the important machine learning algorithms  

Machine learning algorithms

Machine learning algorithms are very much diverse that they can be oriented into various objectives and goals for which machine learning methods are frequently adopted

  • One rule, zero rule, and cubist
  • RIPPER or Repeated Incremental Pruning to Produce Error Reduction
  • Random forest, boosting, and AdaBoost
  • Gradient Boosted Regression Trees and the Stacked Generalization
  • Gradient Boosting Machines and Bootstrapped Aggregation
  • Convolutional Neural Networks and Stacked Autoencoders
  • Deep Boltzmann Machine and Deep Belief Networks
  • Projection Pursuit and Sammon Mapping
  • Principal Component Analysis and Partial Least Square Discriminant Analysis
  • Quadratic Discriminant Analysis and Flexible Discriminant Analysis
  • Partial Least Squares Regression and Multidimensional Scaling
  • Principal Component Regression and Mixture Discriminant Analysis
  • Regularized Discriminant Analysis and Linear Discriminant Analysis
  • K means and K medians
  • Expectation Maximization and Hierarchical Clustering
  • Ridge Regression and Elastic Net
  • Least Angle Regression and the LASSO or Least Absolute Shrinkage and Selection Operator
  • Hopfield Network and perception
  • Black Propagation and Radian Basis Function Network
  • Naive Bayes and Bayesian Network
  • Averaged One Dependents Estimators and Gaussian Naive Bayes
  • Bayesian Belief Networks and Multinomial Naive Bayes
  • Logistic, stepwise, and linear regression
  • Locally Estimated Scatterplot Smoothing and Ordinary Least Squares Regression
  • Multivariate Adaptive Regression Splines
  • MS, C 4.5, C 5.0, and Decision stump
  • Conditional Decision Trees and Iterative Dichotomiser 3
  • Chi-squared Automatic Interaction Detection
  • Classification and regression tree
  • K Nearest Neighbour and Self Organising Map
  • Locally Weighted Learning and Learning Vector Quantization

You can get a complete technical explanation and tips associated with the usage of these algorithms from our website. The selection of your thesis topic for machine learning becomes easier than before when you look into the various aspects of these algorithms and get to choose the best one based on your interests and needs. For this purpose, you can connect with us. We are here to assist you by giving proper expert consultation support for topic selection and allocating a highly qualified team of engineers to carry out your project successfully. Let us now talk about linear regression in detail

What is the process of linear regression?

The following are the three important stages in the process of linear regression analysis

  • Data correlation and directionality analysis
  • Model estimation based on linear fitting
  • Estimation of validity and assessing the merits of the model

It is important that certain characteristic features are inherent in a model for the proper working of an algorithm. Feature engineering is the process by which essential features from raw data are obtained for the better functioning of an algorithm. With the most appropriate features extracted the algorithms become simple. Thus as a result accuracy of results is obtained even in the case of nonideal algorithms. What are the objectives of feature engineering?

  • Preparation of input data for Better compatibility with the chosen machine learning algorithm
  • Enhancement of the efficiency and working of machine learning models

With these goals, feature engineering becomes one of the important aspects of a machine learning research project. Talk to engineers for more details on the methods and algorithms used in extracting the necessary features.  What are the techniques used in feature engineering? 

  • Imputation and binning
  • Log transform and feature split
  • Outliers handling and grouping functions
  • One hot encoding and scaling
  • Data extraction

Usually, we provide practical explanations in easy to understand words to our customers so that all their doubts are cleared even before they start their research. For this purpose, we make use of the real-time implemented models and our successful projects . Check out our website for all our machine learning project details. Let us now talk about hybrid machine learning models.

HYBRID MACHINE LEARNING MODELS

  • When the machine learning methods are integrated with other methods such as optimization approaches, soft computing, and so on drastic improvement can be observed in the machine learning model.
  • The ensemble methods are the culmination of grouping methods like boosting and bagging in case of multiple machine learning classifiers.

Our experts claim that the success of machine learning is dependent on ensemble and hybrid methods advancements. In this regard let us have a look into some of the hybrid methods below

  • NBTree and functional tree
  • Hybrid fuzzy with decision tree
  • Logistic model tree and hybrid hoeffding tree

Most importantly these hybrid models and ensemble-based approaches in machine learning are on a rising scale and our technical team always stays updated about such novelties. So we are highly capable of providing you with the best support in all thesis topics for machine learning. Let us now look into the metrics used in analyzing the performance of machine learning models

Performance analysis of machine learning

Confusion metrics are prominently used for analyzing the machine learning models. The following are the fundamental terms associated with machine learning confusion metrics

  • Contradiction of actual and predicted classes
  • Correct prediction of negative values consisting of ‘no’ results for both actual and predicted classes
  • Correct prediction of positive values consisting of ‘yes’ results for both actual and prediction classes

Using these fundamental parameters the essential values for calculation of efficiency and performance of the machine learning models are obtained as follows.

  • Procession is considered as the ratio between the number of accurate positives predicted and the total number of positives claimed
  • Recall is the ratio of all The true positive rate (in actual class being yes)
  • F1 Score is the average between recall and precision hence taking into account all the false positives and false negatives
  • Uneven distribution of classes require F1 Score to be evaluated than the accuracy, about which we will discuss below
  • Accuracy can be considered in cases of similar false positives and false negatives.
  • For different cost values of false positives and false negatives, it is recommended that you choose to recall and precision for performance evaluation
  • Accuracy is the ratio between correct predictions and the total observations
  • Also accuracy is considered as one of the most important and intuitive measures for analyzing the machine learning system performance

It becomes significant to note here that at thesis topics for machine learning ; our experts have produced excellent results in all these performance metrics. Contact our experts’ team for more details on the approaches that are considered to produce such the best outcomes . We work 24/7 to assist you.

machine learning thesis topics

Opening Hours

  • Mon-Sat 09.00 am – 6.30 pm
  • Lunch Time 12.30 pm – 01.30 pm
  • Break Time 04.00 pm – 04.30 pm
  • 18 years service excellence
  • 40+ country reach
  • 36+ university mou
  • 194+ college mou
  • 6000+ happy customers
  • 100+ employees
  • 240+ writers
  • 60+ developers
  • 45+ researchers
  • 540+ Journal tieup

Payment Options

money gram

Our Clients

machine learning thesis topics

Social Links

machine learning thesis topics

  • Terms of Use

machine learning thesis topics

Opening Time

machine learning thesis topics

Closing Time

  • We follow Indian time zone

award1

machine learning thesis topics

  • Current Members
  • Off-Campus Students
  • Robot Videos
  • Funded Projects
  • Publications by Year
  • Publications by Type
  • Robot Learning Lecture
  • Robot Learning IP
  • Humanoid Robotics Seminar
  • Research Oberseminar
  • New, Open Topics
  • Ongoing Theses
  • Completed Theses
  • External Theses
  • Advice for Thesis Students
  • Thesis Checklist and Template
  • Jobs and Open Positions
  • Current Openings
  • Information for Applicants
  • Application Website
  • TU Darmstadt Student Hiwi Jobs
  • Contact Information

Currently Open Theses Topics

We offer these current topics directly for Bachelor and Master students at TU Darmstadt who can feel free to DIRECTLY contact the thesis advisor if you are interested in one of these topics. Excellent external students from another university may be accepted but are required to first email Jan Peters before contacting any other lab member for a thesis topic. Note that we cannot provide funding for any of these theses projects.

We highly recommend that you do either our robotics and machine learning lectures ( Robot Learning , Statistical Machine Learning ) or our colleagues ( Grundlagen der Robotik , Probabilistic Graphical Models and/or Deep Learning). Even more important to us is that you take both Robot Learning: Integrated Project, Part 1 (Literature Review and Simulation Studies) and Part 2 (Evaluation and Submission to a Conference) before doing a thesis with us.

In addition, we are usually happy to devise new topics on request to suit the abilities of excellent students. Please DIRECTLY contact the thesis advisor if you are interested in one of these topics. When you contact the advisor, it would be nice if you could mention (1) WHY you are interested in the topic (dreams, parts of the problem, etc), and (2) WHAT makes you special for the projects (e.g., class work, project experience, special programming or math skills, prior work, etc.). Supplementary materials (CV, grades, etc) are highly appreciated. Of course, such materials are not mandatory but they help the advisor to see whether the topic is too easy, just about right or too hard for you.

Only contact *ONE* potential advisor at the same time! If you contact a second one without first concluding discussions with the first advisor (i.e., decide for or against the thesis with her or him), we may not consider you at all. Only if you are super excited for at most two topics send an email to both supervisors, so that the supervisors are aware of the additional interest.

FOR FB16+FB18 STUDENTS: Students from other depts at TU Darmstadt (e.g., ME, EE, IST), you need an additional formal supervisor who officially issues the topic. Please do not try to arrange your home dept advisor by yourself but let the supervising IAS member get in touch with that person instead. Multiple professors from other depts have complained that they were asked to co-supervise before getting contacted by our advising lab member.

NEW THESES START HERE

Walk your network: investigating neural network’s location in Q-learning methods.

Scope: Master thesis Advisor: Theo Vincent and Boris Belousov Start: Flexible Topic:

Q-learning methods are at the heart of Reinforcement Learning. They have been shown to outperform humans on some complex tasks such as playing video games [1]. In robotics, where the action space is in most cases continuous, actor-critic methods are relying on Q-learning methods to learn the critic [2]. Although Q-learning methods have been extensively studied in the past, little focus has been placed on the way the online neural network is exploring the space of Q functions. Most approaches focus on crafting a loss that would make the agent learn better policies [3]. Here, we offer a thesis that focuses on the position of the online Q neural network in the space of Q functions. The student will first investigate this idea on simple problems before comparing the performance to strong baselines such as DQN or REM [1, 4] on Atari games. Depending on the result, the student might as well get into MuJoCo and compare the results with SAC [2]. The student will be welcome to propose some ideas as well.

Highly motivated students can apply by sending an email to [email protected] . Please attach your CV and clearly state why you are interested in this topic.

Requirements

  • Strong Python programming skills
  • Knowledge in Reinforcement Learning
  • Experience with deep learning libraries is a plus

References [1] Mnih, Volodymyr, et al. "Human-level control through deep reinforcement learning." nature 518.7540 (2015): 529-533. [2] Haarnoja, Tuomas, et al. "Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor." International conference on machine learning. PMLR, 2018. [3] Hessel, Matteo, et al. "Rainbow: Combining improvements in deep reinforcement learning." Proceedings of the AAAI conference on artificial intelligence. Vol. 32. No. 1. 2018. [4] Agarwal, R., Schuurmans, D. & Norouzi, M.. (2020). An Optimistic Perspective on Offline Reinforcement Learning International Conference on Machine Learning (ICML).

Co-optimizing Hand and Action for Robotic Grasping of Deformable objects

machine learning thesis topics

This project aims to advance deformable object manipulation by co-optimizing robot gripper morphology and control policies. The project will involve utilizing existing simulation environments for deformable object manipulation [2] and implementing a method to jointly optimize gripper morphology and grasp policies within the simulation.

Required Qualification:

  • Familiarity with deep learning libraries such as PyTorch or Tensorflow

Preferred Qualification:

  • Attendance of the lectures "Statistical Machine Learning", "Computational Engineering and Robotics" and "Robot Learning"

Application Requirements:

  • Curriculum Vitae
  • Motivation letter explaining why you would like to work on this topic and why you are the perfect candidate

Interested students can apply by sending an e-mail to [email protected] and attaching the required documents mentioned above.

References: [1] Xu, Jie, et al. "An End-to-End Differentiable Framework for Contact-Aware Robot Design." Robotics: Science & Systems. 2021. [2] Huang, Isabella, et al. "DefGraspNets: Grasp Planning on 3D Fields with Graph Neural Nets." arXiv preprint arXiv:2303.16138 (2023).

Geometry-Aware Diffusion Models for Robotics

In this thesis, you will work on developing an imitation learning algorithm using diffusion models for robotic manipulation tasks, such as the ones in [2, 3, 4], but taking into account the geometry of the task space.

If this sounds interesting, please send an email to [email protected] and [email protected] , and possibly attach your CV, highlighting the relevant courses you took in robotics and machine learning.

What's in it for you:

  • You get to work on an exciting topic at the intersection of deep-learning and robotics
  • We will supervise you closely throughout your thesis
  • Depending on the results, we will aim for an international conference publication

Requirements:

  • Be motivated -- we will support you a lot, but we expect you to contribute a lot too
  • Robotics knowledge
  • Experience setting up deep learning pipelines -- from data collection, architecture design, training, and evaluation
  • PyTorch -- especially experience writing good parallelizable code (i.e., runs fast in the GPU)

References: [1] https://arxiv.org/abs/2112.10752 [2] https://arxiv.org/abs/2308.01557 [3] https://arxiv.org/abs/2209.03855 [4] https://arxiv.org/abs/2303.04137 [5] https://arxiv.org/abs/2205.09991

Learning Latent Representations for Embodied Agents

machine learning thesis topics

Interested students can apply by sending an E-Mail to [email protected] and attaching the required documents mentioned below.

  • Experience with TensorFlow/PyTorch
  • Familiarity with core Machine Learning topics
  • Experience programming/controlling robots (either simulated or real world)
  • Knowledgeable about different robot platforms (quadrupeds and bipedal robots)
  • Resume / CV
  • Cover letter explaining why this topic fits you well and why you are an ideal candidate

References: [1] Ho and Ermon. "Generative adversarial imitation learning" [2] Arenz, et al. "Efficient Gradient-Free Variational Inference using Policy Search"

Characterizing Fear-induced Adaptation of Balance by Inverse Reinforcement Learning

machine learning thesis topics

Interested students can apply by sending an E-Mail to [email protected] and attaching the required documents mentioned below.

  • Basic knowledge of reinforcement learning
  • Hand-on experience with reinforcement learning or inverse reinforcement learning
  • Cognitive science background

References: [1] Maki, et al. "Fear of Falling and Postural Performance in the Elderly" [2] Davis et al. "The relationship between fear of falling and human postural control" [3] Ho and Ermon. "Generative adversarial imitation learning"

Timing is Key: CPGs for regularizing Quadruped Gaits learned with DRL

To tackle this problem we want to utilize Central Pattern Generators (CPGs), which can generate timings for ground contacts for the four feet. The policy gets rewarded for complying with the contact patterns of the CPGs. This leads to a straightforward way of regularizing and steering the policy to a natural gait without posing too strong restrictions on it. We first want to manually find fitting CPG parameters for different gait velocities and later move to learning those parameters in an end-to-end fashion.

Highly motivated students can apply by sending an E-Mail to [email protected] and attaching the required documents mentioned below.

Minimum Qualification:

  • Good Python programming skills
  • Basic knowledge of the PyTorch library
  • Basic knowledge of Reinforcement Learning
  • Good knowledge of the PyTorch library
  • Basic knowledge of the MuJoCo simulator

References: [1] Cheng, Xuxin, et al. "Extreme Parkour with Legged Robots."

Damage-aware Reinforcement Learning for Deformable and Fragile Objects

machine learning thesis topics

Goal of this thesis will be the development and application of a model-based reinforcement learning method on real robots. Your tasks will include: 1. Setting up a simulation environment for deformable object manipulation 2. Utilizing existing models for stress and deformability prediction[1] 3. Implementing a reinforcement learning method to work in simulation and, if possible, on the real robot methods.

If you are interested in this thesis topic and believe you possess the necessary skills and qualifications, please submit your application, including a resume and a brief motivation letter explaining your interest and relevant experience. Please send your application to [email protected].

Required Qualification :

  • Enthusiasm for and experience in robotics, machine learning, and simulation
  • Strong programming skills in Python

Desired Qualification :

  • Attendance of the lectures "Statistical Machine Learning", "Computational Engineering and Robotics" and (optionally) "Robot Learning"

References: [1] Huang, I., Narang, Y., Bajcsy, R., Ramos, F., Hermans, T., & Fox, D. (2023). DefGraspNets: Grasp Planning on 3D Fields with Graph Neural Nets. arXiv preprint arXiv:2303.16138.

Imitation Learning meets Diffusion Models for Robotics

machine learning thesis topics

The objective of this thesis is to build upon prior research [2, 3] to establish a connection between Diffusion Models and Imitation Learning. We aim to explore how to exploit Diffusion Models and improve the performance of Imitation learning algorithms that interact with the world.

We welcome highly motivated students to apply for this opportunity by sending an email expressing their interest to Firas Al-Hafez ( [email protected] ) Julen Urain ( [email protected] ). Please attach your letter of motivation and CV, and clearly state why you are interested in this topic and why you are the ideal candidate for this position.

Required Qualification : 1. Strong Python programming skills 2. Basic Knowledge in Imitation Learning 3. Interest in Diffusion models, Reinforcement Learning

Desired Qualification : 1. Attendance of the lectures "Statistical Machine Learning", "Computational Engineering and Robotics" and/or "Reinforcement Learning: From Fundamentals to the Deep Approaches"

References: [1] Song, Yang, and Stefano Ermon. "Generative modeling by estimating gradients of the data distribution." Advances in neural information processing systems 32 (2019). [2] Ho, Jonathan, and Stefano Ermon. "Generative adversarial imitation learning." Advances in neural information processing systems 29 (2016). [3] Garg, D., Chakraborty, S., Cundy, C., Song, J., & Ermon, S. (2021). Iq-learn: Inverse soft-q learning for imitation. Advances in Neural Information Processing Systems, 34, 4028-4039. [4] Chen, R. T., & Lipman, Y. (2023). Riemannian flow matching on general geometries. arXiv preprint arXiv:2302.03660.

The Power of Friendship: Dynamic Multi-Agent Reward Sharing

machine learning thesis topics

In this thesis, we want to investigate the use of generalised reward-sharing between agents that would allow for collaboration in a less constrained manner than previous methods. The method would assume partial observability of each agent, and a bi-level optimisation system that would disseminate scaled rewards between agents in addition to environmental rewards. We aim to demonstrate good performance on small-scale multi-agent settings (StarCraft, Prisoner's Dilemma, Stag Hunt) and to outperform existing methods in large-scale or dynamic settings (Braess's Paradox, network optimisation).

How to apply

  • Send an email to [email protected] with your CV and a short motivation email explaining your reason for applying for this thesis and your academic/career objectives.

Minimum knowledge

  • Good Python programming skills;
  • Basic knowledge of Reinforcement Learning.

Preferred knowledge

  • Knowledge of the PyTorch library;
  • Knowledge of the MushroomRL library.
  • Get to work on a difficult open problem in deep learning with important mathematical underpinnings
  • Possible publication at an international conference (depending on results)

References: [1] Peysakhovich, Alexander, and Adam Lerer. "Prosocial learning agents solve generalized stag hunts better than selfish ones." arXiv preprint arXiv:1709.02865 (2017). [2] Jaques, Natasha, et al. "Social influence as intrinsic motivation for multi-agent deep reinforcement learning." International conference on machine learning. PMLR, 2019. [3] Yi, Yuxuan, et al. "Learning to Share in Networked Multi-Agent Reinforcement Learning." Advances in Neural Information Processing Systems 35 (2022): 15119-15131.

  • Be extremely motivated -- we will support you a lot, but we expect you to contribute a lot too

Scaling Behavior Cloning to Humanoid Locomotion

Scope: Bachelor / Master thesis Advisor: Joe Watson Added: 2023-10-07 Start: ASAP Topic: In a previous project [1], I found that behavior cloning (BC) was a surprisingly poor baseline for imitating humanoid locomotion. I suspect the issue may lie in the challenges of regularizing high-dimensional regression.

The goal of this project is to investigate BC for humanoid imitation, understand the scaling issues present, and evaluate possible solutions, e.g. regularization strategies from the regression literature.

The project will be building off Google Deepmind's Acme library [2], which has BC algorithms and humanoid demonstration datasets [3] already implemented, and will serve as the foundation of the project.

To apply, email [email protected] , ideally with a CV and transcript so I can assess your suitability.

  • Experience, interest and enthusiasm for the intersection of robot learning and machine learning
  • Experience with Acme and JAX would be a benefit, but not necessary

References: [1] https://arxiv.org/abs/2305.16498 [2] https://github.com/google-deepmind/acme [3] https://arxiv.org/abs/2106.00672

Robot Gaze for Communicating Collision Avoidance Intent in Shared Workspaces

Scope: Bachelor/Master thesis Advisor: Alap Kshirsagar , Dorothea Koert Added: 2023-09-27 Start: ASAP

machine learning thesis topics

Topic: In order to operate close to non-experts, future robots require both an intuitive form of instruction accessible to lay users and the ability to react appropriately to a human co-worker. Instruction by imitation learning with probabilistic movement primitives (ProMPs) [1] allows capturing tasks by learning robot trajectories from demonstrations including the motion variability. However, appropriate responses to human co-workers during the execution of the learned movements are crucial for fluent task execution, perceived safety, and subjective comfort. To facilitate such appropriate responsive behaviors in human-robot interaction, the robot needs to be able to react to its human workspace co-inhabitant online during the execution. Also, the robot needs to communicate its motion intent to the human through non-verbal gestures such as eye and head gazes [2][3]. In particular for humanoid robots, combining motions of arms with expressive head and gaze directions is a promising approach that has not yet been extensively studied in related work.

Goals of the thesis:

  • Develop a method to combine robot head/gaze motion with ProMPs for online collision avoidance
  • Implement the method on a Franka-Emika Panda Robot
  • Evaluate and compare the implemented behaviors in a study with human participants

Highly motivated students can apply by sending an email to [email protected]. Please attach your CV and transcript, and clearly state your prior experiences and why you are interested in this topic.

  • Strong Programming Skills in python
  • Prior experience with Robot Operating System (ROS) and user studies would be beneficial
  • Strong motivation for human-centered robotics including design and implementation of a user study

References : [1] Koert, Dorothea, et al. "Learning intention aware online adaptation of movement primitives." IEEE Robotics and Automation Letters 4.4 (2019): 3719-3726. [2] Admoni, Henny, and Brian Scassellati. "Social eye gaze in human-robot interaction: a review." Journal of Human-Robot Interaction 6.1 (2017): 25-63. [3] Lemasurier, Gregory, et al. "Methods for expressing robot intent for human–robot collaboration in shared workspaces." ACM Transactions on Human-Robot Interaction (THRI) 10.4 (2021): 1-27.

Tactile Sensing for the Real World

Topic: Tactile sensing is a crucial sensing modality that allows humans to perform dexterous manipulation[1]. In recent years, the development of artificial tactile sensors has made substantial progress, with current models relying on cameras inside the fingertips to extract information about the points of contact [2]. However, robotic tactile sensing is still a largely unsolved topic despite these developments. A central challenge of tactile sensing is the extraction of usable representations of sensor readings, especially since these generally contain an incomplete view of the environment.

Recent model-based reinforcement learning methods like Dreamer [3] leverage latent state-space models to reason about the environment from partial and noisy observations. However, more work has yet to be done to apply such methods to real-world manipulation tasks. Hence, this thesis will explore whether Dreamer can solve challenging real-world manipulation tasks by leveraging tactile information. Initial results suggest that tasks like peg-in-a-hole can indeed be solved with Dreamer in simulation (see figure above), but the applicability of this method in the real world has yet to be shown.

In this work, you will work with state-of-the-art hardware and compute resources on a hot research topic with the option of publishing your work at a scientific conference.

Highly motivated students can apply by sending an email to [email protected]. Please attach a transcript of records and clearly state your prior experiences and why you are interested in this topic.

  • Ideally experience with deep learning libraries like JAX or PyTorch
  • Experience with reinforcement learning is a plus
  • Experience with Linux

References [1] 2S Match Anest2, Roland Johansson Lab (2005), https://www.youtube.com/watch?v=HH6QD0MgqDQ [2] Gelsight Inc., Gelsight Mini, https://www.gelsight.com/gelsightmini/ [3] Hafner, D., Lillicrap, T., Ba, J., & Norouzi, M. (2019). Dream to control: Learning behaviors by latent imagination. arXiv preprint arXiv:1912.01603.

Large Vision-Language Neural Networks for Open-Vocabulary Robotic Manipulation

machine learning thesis topics

Robots are expected to soon leave their factory/laboratory enclosures and operate autonomously in everyday unstructured environments such as households. Semantic information is especially important when considering real-world robotic applications where the robot needs to re-arrange objects as per a set of language instructions or human inputs (as shown in the figure). Many sophisticated semantic segmentation networks exist [1]. However, a challenge when using such methods in the real world is that the semantic classes rarely align perfectly with the language input received by the robot. For instance, a human language instruction might request a ‘glass’ or ‘water’, but the semantic classes detected might be ‘cup’ or ‘drink’.

Nevertheless, with the rise of large language and vision-language models, we now have capable segmentation models that do not directly predict semantic classes but use learned associations between language queries and classes to give us ’open-vocabulary’ segmentation [2]. Some models are especially powerful since they can be used with arbitrary language queries.

In this thesis, we aim to build on advances in 3D vision-based robot manipulation and large open-vocabulary vision models [2] to build a full pick-and-place pipeline for real-world manipulation. We also aim to find synergies between scene reconstruction and semantic segmentation to determine if knowing the object semantics can aid the reconstruction of the objects and, in turn, aid manipulation.

Highly motivated students can apply by sending an e-mail expressing their interest to Snehal Jauhri (email: [email protected]) or Ali Younes (email: [email protected]), attaching your letter of motivation and possibly your CV.

Topic in detail : Thesis_Doc.pdf

Requirements: Enthusiasm, ambition, and a curious mind go a long way. There will be ample supervision provided to help the student understand basic as well as advanced concepts. However, prior knowledge of computer vision, robotics, and Python programming would be a plus.

References: [1] Y. Wu, A. Kirillov, F. Massa, W.-Y. Lo, and R. Girshick, “Detectron2”, https://github.com/facebookresearch/detectron2 , 2019. [2] F. Liang, B. Wu, X. Dai, K. Li, Y. Zhao, H. Zhang, P. Zhang, P. Vajda, and D. Marculescu, “Open-vocabulary semantic segmentation with mask-adapted clip,” in CVPR, 2023, pp. 7061–7070, https://github.com/facebookresearch/ov-seg

Dynamic Tiles for Deep Reinforcement Learning

machine learning thesis topics

Linear approximators in Reinforcement Learning are well-studied and come with an in-depth theoretical analysis. However, linear methods require defining a set of features of the state to be used by the linear approximation. Unfortunately, the feature construction process is a particularly problematic and challenging task. Deep Reinforcement learning methods have been introduced to mitigate the feature construction problem: these methods do not require handcrafted features, as features are extracted automatically by the network during learning, using gradient descent techniques.

In simple reinforcement learning tasks, however, it is possible to use tile coding as features: Tiles are simply a convenient discretization of the state space that allows us to easily control the generalization capabilities of the linear approximator. The objective of this thesis is to design a novel algorithm for automatic feature extraction that generates a set of features similar to tile coding, but that can arbitrarily partition the state space and deal with arbitrary complex state space, such as images. The idea is to combine the feature extraction problem directly with Linear Reinforcement Learning methods, defining an algorithm that is able both to have the theoretical guarantees and good convergence properties of these methods and the flexibility of Deep Learning approaches.

  • Curriculum Vitae (CV);
  • A motivation letter explaining the reason for applying for this thesis and academic/career objectives.
  • Knowledge of the Atari environments (ale-py library).

Accepted candidate will

  • Define a generalization of tile coding working with an arbitrary input set (including images);
  • Design a learning algorithm to adapt the tiles using data of interaction with the environment;
  • Combine feature learning with standard linear methods for Reinforcement Learning;
  • Verify the novel methodology in simple continuous state and discrete actions environments;
  • (Optionally) Extend the experimental analysis to the Atari environment setting.

Deep Learning Meets Teleoperation: Constructing Learnable and Stable Inductive Guidance for Shared Control

This work considers policies as learnable inductive guidance for shared control. In particular, we use the class of Riemannian motion policies [3] and consider them as differentiable optimization layers [4]. We analyze (i) if RMPs can be pre-trained by learning from demonstrations [5] or reinforcement learning [6] given a specific context; (ii) and subsequently employed seamlessly for human-guided teleoperation thanks to their physically consistent properties, such as stability [3]. We believe this step eliminates the laborious process of constructing complex policies and leads to improved and generalizable shared control architectures.

Highly motivated students can apply by sending an e-mail expressing your interest to [email protected] and [email protected] , attaching your letter of motivation and possibly your CV.

  • Experience with deep learning libraries (in particular Pytorch)
  • Knowledge in reinforcement learning and/or machine learning

References: [1] Niemeyer, Günter, et al. "Telerobotics." Springer handbook of robotics (2016); [2] Selvaggio, Mario, et al. "Autonomy in physical human-robot interaction: A brief survey." IEEE RAL (2021); [3] Cheng, Ching-An, et al. "RMP flow: A Computational Graph for Automatic Motion Policy Generation." Springer (2020); [4] Jaquier, Noémie, et al. "Learning to sequence and blend robot skills via differentiable optimization." IEEE RAL (2022); [5] Mukadam, Mustafa, et al. "Riemannian motion policy fusion through learnable lyapunov function reshaping." CoRL (2020); [6] Xie, Mandy, et al. "Neural geometric fabrics: Efficiently learning high-dimensional policies from demonstration." CoRL (2023).

Dynamic symphony: Seamless human-robot collaboration through hierarchical policy blending

This work focuses on arbitration between the user and assistive policy, i.e., shared autonomy. Various works allow the user to influence the dynamic behavior explicitly and, therefore, could not satisfy stability guarantees [3]. We pursue the idea of formulating arbitration as a trajectory-tracking problem that implicitly considers the user's desired behavior as an objective [4]. Therefore, we extend the work of Hansel et al. [5], who employed probabilistic inference for policy blending in robot motion control. The proposed method corresponds to a sampling-based online planner that superposes reactive policies given a predefined objective. This method enables the user to implicitly influence the behavior without injecting energy into the system, thus satisfying stability properties. We believe this step leads to an alternative view of shared autonomy with an improved and generalizable framework.

Highly motivated students can apply by sending an e-mail expressing your interest to [email protected] or [email protected] , attaching your letter of motivation and possibly your CV.

References: [1] Niemeyer, Günter, et al. "Telerobotics." Springer handbook of robotics (2016); [2] Selvaggio, Mario, et al. "Autonomy in physical human-robot interaction: A brief survey." IEEE RAL (2021); [3] Dragan, Anca D., and Siddhartha S. Srinivasa. "A policy-blending formalism for shared control." IJRR (2013); [4] Javdani, Shervin, et al. "Shared autonomy via hindsight optimization for teleoperation and teaming." IJRR (2018); [5] Hansel, Kay, et al. "Hierarchical Policy Blending as Inference for Reactive Robot Control." IEEE ICRA (2023).

Feeling the Heat: Igniting Matches via Tactile Sensing and Human Demonstrations

In this thesis, we want to investigate the effectiveness of vision-based tactile sensors for solving dynamic tasks (igniting matches). Since the whole task is difficult to simulate, we directly collect real-world data to learn policies from the human demonstrations [2,3]. We believe that this work is an important step towards more advanced tactile skills.

Highly motivated students can apply by sending an e-mail expressing your interest to [email protected] and [email protected] , attaching your letter of motivation and possibly your CV.

  • Good knowledge of Python
  • Prior experience with real robots and Linux is a plus

References: [1] https://www.youtube.com/watch?v=HH6QD0MgqDQ [2] Learning Compliant Manipulation through Kinesthetic and Tactile Human-Robot Interaction; Klas Kronander and Aude Billard. [3] https://www.youtube.com/watch?v=jAtNvfPrKH8

Inverse Reinforcement Learning for Neuromuscular Control of Humanoids

Within this thesis, the problems of learning from observations and efficient exploration in overactued systems should be addressed. Regarding the former, novel methods incorporating inverse dynamics models into the inverse reinforcement learning problem [1] should be adapted and applied. To address the problem of efficient exploration in overactuted systems, two approaches should be implemented and compared. The first approach uses a handcrafted action space, which disables and modulates actions in different phases of the gait based on biomechanics knowledge [2]. The second approach uses a stateful policy to incorporate an inductive bias into the policy [3]. The thesis will be supervised in conjunction with Guoping Zhao ( [email protected] ) from the locomotion lab.

Highly motivated students can apply by sending an e-mail expressing their interest to Firas Al-Hafez ( [email protected] ), attaching your letter of motivation and possibly your CV. Try to make clear why you would like to work on this topic, and why you would be the perfect candidate for the latter.

Required Qualification : 1. Strong Python programming skills 2. Knowledge in Reinforcement Learning 3. Interest in understanding human locomotion

Desired Qualification : 1. Hands-on experience on robotics-related RL projects 2. Prior experience with different simulators 3. Attendance of the lectures "Statistical Machine Learning", "Computational Engineering and Robotics" and/or "Reinforcement Learning: From Fundamentals to the Deep Approaches"

References: [1] Al-Hafez, F.; Tateo, D.; Arenz, O.; Zhao, G.; Peters, J. (2023). LS-IQ: Implicit Reward Regularization for Inverse Reinforcement Learning, International Conference on Learning Representations (ICLR). [2] Ong CF; Geijtenbeek T.; Hicks JL; Delp SL (2019) Predicting gait adaptations due to ankle plantarflexor muscle weakness and contracture using physics-based musculoskeletal simulations. PLoS Computational Biology [3] Srouji, M.; Zhang, J:;Salakhutdinow, R. (2018) Structured Control Nets for Deep Reinforcement Learning, International Conference on Machine Learning (ICML)

Robotic Tactile Exploratory Procedures for Identifying Object Properties

machine learning thesis topics

Goals of the thesis

  • Literature review of robotic EPs for identifying object properties [2,3,4]
  • Develop and implement robotic EPs for a Digit tactile sensor
  • Compare performance of robotic EPs with human EPs

Desired Qualifications

  • Interested in working with real robotic systems
  • Python programming skills

Literature [1] Lederman and Klatzky, “Haptic perception: a tutorial” [2] Seminara et al., “Active Haptic Perception in Robots: A Review” [3] Chu et al., “Using robotic exploratory procedures to learn the meaning of haptic adjectives” [4] Kerzel et al., “Neuro-Robotic Haptic Object Classification by Active Exploration on a Novel Dataset”

Scaling learned, graph-based assembly policies

machine learning thesis topics

  • scaling our previous methods to incorporate mobile manipulators or the Kobo bi-manual manipulation platform. The increased workspace of both would allow for handling a wider range of objects
  • [2] has shown more powerful, yet, it includes running a MILP for every desired structure. Thus another idea could be to investigate approaches aiming to approximate this solution
  • adapting the methods to handle more irregular-shaped objects / investigate curriculum learning

Highly motivated students can apply by sending an e-mail expressing your interest to [email protected] , attaching your letter of motivation and possibly your CV.

  • Experience with deep learning libraries (in particular Pytorch) is a plus
  • Experience with reinforcement learning / having taken Robot Learning is also a plus

References: [1] Learn2Assemble with Structured Representations and Search for Robotic Architectural Construction; Niklas Funk et al. [2] Graph-based Reinforcement Learning meets Mixed Integer Programs: An application to 3D robot assembly discovery; Niklas Funk et al. [3] Structured agents for physical construction; Victor Bapst et al.

Long-Horizon Manipulation Tasks from Visual Imitation Learning (LHMT-VIL): Algorithm

machine learning thesis topics

The proposed architecture can be broken down into the following sub-tasks: 1. Multi-object 6D pose estimation from video: Identify the object 6D poses in each video frame to generate the object trajectories 2. Action segmentation from video: Classify the action being performed in each video frame 3. High-level task representation learning: Learn the sequence of robotic movement primitives with the associated object poses such that the robot completes the demonstrated task 4. Low-level movement primitives: Create a database of low-level robotic movement primitives which can be sequenced to solve the long-horizon task

Desired Qualification: 1. Strong Python programming skills 2. Prior experience in Computer Vision and/or Robotics is preferred

Long-Horizon Manipulation Tasks from Visual Imitation Learning (LHMT-VIL): Dataset

During the project, we will create a large-scale dataset of videos of humans demonstrating industrial assembly sequences. The dataset will contain information of the 6D poses of the objects, the hand and body poses of the human, the action sequences among numerous other features. The dataset will be open-sourced to encourage further research on VIL.

[1] F. Sener, et al. "Assembly101: A Large-Scale Multi-View Video Dataset for Understanding Procedural Activities". CVPR 2022. [2] P. Sharma, et al. "Multiple Interactions Made Easy (MIME) : Large Scale Demonstrations Data for Imitation." CoRL, 2018.

Adaptive Human-Robot Interactions with Human Trust Maximization

machine learning thesis topics

  • Good knowledge of Python and/or C++;
  • Good knowledge in Robotics and Machine Learning;
  • Good knowledge of Deep Learning frameworks, e.g, PyTorch;

References: [1] Xu, Anqi, and Gregory Dudek. "Optimo: Online probabilistic trust inference model for asymmetric human-robot collaborations." ACM/IEEE HRI, IEEE, 2015; [2] Kwon, Minae, et al. "When humans aren’t optimal: Robots that collaborate with risk-aware humans." ACM/IEEE HRI, IEEE, 2020; [3] Chen, Min, et al. "Planning with trust for human-robot collaboration." ACM/IEEE HRI, IEEE, 2018; [4] Poole, Ben et al. “On variational bounds of mutual information”. ICML, PMLR, 2019.

Causal inference of human behavior dynamics for physical Human-Robot Interactions

machine learning thesis topics

Highly motivated students can apply by sending an e-mail expressing your interest to [email protected] , attaching your a letter of motivation and possibly your CV.

  • Good knowledge of Robotics;
  • Good knowledge of Deep Learning frameworks, e.g, PyTorch
  • Li, Q., Chalvatzaki, G., Peters, J., Wang, Y., Directed Acyclic Graph Neural Network for Human Motion Prediction, 2021 IEEE International Conference on Robotics and Automation (ICRA).
  • Löwe, S., Madras, D., Zemel, R. and Welling, M., 2020. Amortized causal discovery: Learning to infer causal graphs from time-series data. arXiv preprint arXiv:2006.10833.
  • Yang, W., Paxton, C., Mousavian, A., Chao, Y.W., Cakmak, M. and Fox, D., 2020. Reactive human-to-robot handovers of arbitrary objects. arXiv preprint arXiv:2011.08961.

Incorporating First and Second Order Mental Models for Human-Robot Cooperative Manipulation Under Partial Observability

Scope: Master Thesis Advisor: Dorothea Koert , Joni Pajarinen Added: 2021-06-08 Start: ASAP

machine learning thesis topics

The ability to model the beliefs and goals of a partner is an essential part of cooperative tasks. While humans develop theory of mind models for this aim already at a very early age [1] it is still an open question how to implement and make use of such models for cooperative robots [2,3,4]. In particular, in shared workspaces human robot collaboration could potentially profit from the use of such models e.g. if the robot can detect and react to planned human goals or a human's false beliefs during task execution. To make such robots a reality, the goal of this thesis is to investigate the use of first and second order mental models in a cooperative manipulation task under partial observability. Partially observable Markov decision processes (POMDPs) and interactive POMDPs (I-POMDPs) [5] define an optimal solution to the mental modeling task and may provide a solid theoretical basis for modelling. The thesis may also compare related approaches from the literature and setup an experimental design for evaluation with the bi-manual robot platform Kobo.

Highly motivated students can apply by sending an e-mail expressing your interest to [email protected] attaching your CV and transcripts.

References:

  • Wimmer, H., & Perner, J. Beliefs about beliefs: Representation and constraining function of wrong beliefs in young children's understanding of deception (1983)
  • Sandra Devin and Rachid Alami. An implemented theory of mind to improve human-robot shared plans execution (2016)
  • Neil Rabinowitz, Frank Perbet, Francis Song, Chiyuan Zhang, SM Ali Eslami,and Matthew Botvinick. Machine theory of mind (2018)
  • Connor Brooks and Daniel Szafir. Building second-order mental models for human-robot interaction. (2019)
  • Prashant Doshi, Xia Qu, Adam Goodie, and Diana Young. Modeling recursive reasoning by humans using empirically informed interactive pomdps. (2010)

My Paper Done

  • Services Paper editing services Paper proofreading Business papers Philosophy papers Write my paper Term papers for sale Term paper help Academic term papers Buy research papers College writing services Paper writing help Student papers Original term papers Research paper help Nursing papers for sale Psychology papers Economics papers Medical papers Blog

machine learning thesis topics

147 Top Machine Learning Topics To Get Your Paper Easy

machine learning topics

First of all, let’s determine – what is machine learning ? Generally, it refers to the study of computer algorithms that improve automatically through the use of data and experience. Machine learning is seen as part of AI that makes decisions or predictions without being entirely programmed. The complexity of developing conventional algorithms for performing the much-needed tasks makes this field a choice for the chosen few. Statistics show that the number of college students pursuing this course is few. Are you among the chosen few who would like to improve and excel in your computer science course? Well, our expert help will help springboard you to the next level in writing a research paper . 

How To Find Topics in Machine Learning

The task of sourcing for impeccable machine learning topics is the least trodden path. The few resources available in the field and the course’s technicality make it all the more difficult. However, there are places you can find top-tier ideas in machine learning:

Reputable online sources Your well-stocked college library Available computer science papers (articles, journals, theses, etc.) Tech programs such as TED-X

With these readily available sources, you can be confident of a long list of machine learning topics that can impress your professor. Use our online research paper writing servic e and get your winning paper done fast. That said, our professionals have compiled a handpicked list of writing ideas for your inspiration. Have a look at them:

The Best Machine Learning Topics

  • Discuss supervised learning algorithms approach in machine learning
  • Evaluate iterative optimization of an object function
  • How can a function be used to determine the output for inputs correctly?
  • How to improve the accuracy of predictions or outputs with an algorithm
  • A case study of the classification, active learning, and regression in college
  • Analyze the effectiveness of learning from examples using a similarity function
  • The application of unsupervised learning in density estimation in statistics
  • Manifold learning algorithms for solving machine learning problems
  • Discuss the application of machine learning in data mining
  • How to detect anomalies and deviations in machine learning
  • The role of developmental robotics in machine learning
  • Determine the relationship between variables in large databases
  • How to develop artificial immune systems
  • Discuss the concept of strict rules in machine learning
  • Learning classifier systems in machine learning algorithms

Easy Machine Learning Research Topics

  • Challenges involved in creating intelligent machines that mimic human behavior
  • Discuss the process of data observations in machine learning
  • How to enable computers to learn automatically without human intervention
  • How to analyze training data and produce an inferred function
  • Drawing inferences from datasets comprising of input data without labeled responses
  • The role of Artificial Neural Network in learning from observational data
  • Evaluate the input and output layers of artificial neural networks
  • How to stack multiple layers of neural networks to create a huge network
  • Discuss the dependence of machine learning on linear regression
  • Dealing with the classification problem using logistic regression
  • The random forest machine learning technique in college

Hot Topics in Machine Learning

  • New computing technologies that have contributed to machine learning
  • The essence of machine learning in developing the self-driving Google car
  • An analysis of online recommendation offers: A case of Netflix
  • How to know what customers are saying about a product using machine learning
  • The crucial role of fraud detection in machine learning
  • Discuss the crucial relationship between AI and machine learning
  • What has contributed to the resurging interest in machine learning?
  • The role of machine learning in computational processing
  • The impact of machine learning in developing faster and more accurate results
  • The role of data preparation capabilities in machine learning
  • Discuss the place of machine learning in today’s world

Interesting Machine Learning Thesis Topics

  • How to apply machine learning to the progressive Internet of Things
  • Why industries using large amounts of data need machine learning knowledge
  • The role of machine learning in banks and other financial institutions
  • How government agencies use machine learning in ensuring public safety
  • Analyze how sensor data is used in identifying ways to increase efficiency
  • The role of wearable devices to the healthcare industry
  • How website recommending systems are transforming the retail sector
  • The process of finding new energy sources using machine learning
  • How to identify patterns and trends in transportation using machine learning
  • Discuss prediction and gradient boosting as machine learning methods
  • Compare and contrast between machine learning, deep learning, and data mining

Top Machine Learning Project Topics

  • How to pair the best algorithms with the right tools in machine learning
  • The important role of the rich, sophisticated heritage of statistics in machine learning
  • The role of machine learning in huge enterprise environments
  • Discuss some of the local search optimization techniques: A case of genetic algorithms
  • How to handle multivariate adaptive regression splines
  • Discuss the effectiveness of the singular value decomposition
  • Tools and processes involved in machine learning: A case of algorithms
  • Evaluate the process of comprehensive data quality and management
  • The interactive data exploration and visualization model
  • Compare and contrast the different machine learning models today
  • How the automated sensor ensemble model is used in identifying flaws

College Research Topics in Machine Learning

  • How to determine the best machine-learning algorithm to use
  • The role of curiosity in meeting the challenges that lie ahead of machine learning
  • How scientists have incorporated machine learning in combating the pandemic
  • The place of innovation, agility, and customer-centricity in machine learning
  • The underpinnings of resilience in the machine learning process
  • The role of machine learning in the face of unpredictability
  • Top-rated analytical skills gained through machine learning
  • Getting repeatable data using the easy model deployment
  • Discuss the Graphical User Interfaces for building models and process flaws
  • Evaluate the sequential covering rule building
  • Principal component analysis in the machine learning process

Machine Learning Hot Topics

  • Developing a stock price detector using machine learning
  • Discuss how to predict wine quality using a wine quality dataset
  • The process of developing human activity recognition using a smartphone dataset
  • Evaluate object detection with deep learning
  • Why do we need to develop machine learning projects?
  • Why there are a lot of unearthed projects in software development
  • Machine learning: The efficiency of using textbooks and study materials
  • Getting hands-on experience through machine learning
  • Effective software for developing projects in machine learning
  • Why data scientists are going to be the future of the world
  • How to leverage various Artificial Intelligence technologies

Current Research Topics in Machine Learning

  • How to cartoony an image with machine learning
  • The role of machine learning in aiding coronavirus patients
  • How easy is it to classify human facial expressions and map them to emojis?
  • The role of machine learning in the increased cyberbullying claims
  • Why most developing countries are slow to incorporating machine learning
  • The effectiveness of the machine learning curriculum in colleges and universities
  • Are internet sources watering down the essence of machine learning
  • Discuss the role of machine learning in developing bioweapons
  • Using machine learning to solve daily problems in life
  • How effectively can machines recognize handwritten digits?
  • The role of convolutional neural networks in machine learning

Advanced Topics in AI & Machine Learning

  • Discuss the latest generative models in machine learning
  • The role of the Bayesian inference in the mathematics of machine learning
  • How probabilistic programming is transforming machine learning
  • Model selection and learning: The challenges herein
  • Discuss the application of machine learning in natural language processing
  • The development of neural Turing machines
  • Evaluate syntactic and semantic parsing in the process of machine learning
  • Discuss GPU optimization for neural networks
  • Back-propagation of time through machine learning processes
  • The role of MIT in advancing research in machine learning
  • Long-short term memory: A case study of the applications of machine learning

Best Machine Learning Project Topics

  • Advances made in machine learning in the recent years
  • A simple way of preventing neural networks from overfitting
  • How to use deep residual learning for image recognition
  • The process of accelerating deep network training through batch normalization
  • Discuss large-scale video classification with convolutional neural networks
  • Evaluate some of the common objects in Microsoft COCO
  • Describe how to learn deep machine features for scene recognition
  • Developing a new framework for generative adversarial nets
  • The impact of high-speed tracking with kernelled correlation features
  • A review of the multi-label learning algorithms
  • Describe how to transfer features in deep neural networks

Top-Rated Machine Learning Research Project Topics

  • Why we do not have hundreds of classifiers to solve real-world problems of classification
  • A web-scale approach to dealing with probabilistic knowledge
  • Supervised machine learning methods for fusing distinct information sources
  • Suggest new algorithms for evaluating and comparing algorithms
  • A review of the existing trends in extreme learning machines
  • A survey of the concept drift adaptation in machine learning
  • Describe the simultaneous segmentation and detection process
  • Discuss the most used feature selection methods today
  • The problem of Face Alignment for a single image
  • Evaluate the various multiple classifier systems in the world
  • How to achieve a super-real-time performance with high-quality predictions

Credible Machine Learning Dissertation Topics

  • Describe a semi-supervised setting in machine learning
  • Concepts of hypothesis sets in machine learning
  • Preprocessing of data: A case study of data normalization
  • Some of the most common problems in machine learning
  • Terminology and basic concepts: A case study of convex optimization
  • Discuss batch gradient descent and stochastic gradient descent
  • Assess the notion of support vectors in support machines
  • Online tools used for getting some intuition of an algorithm
  • Describe the generative model and basic ideas of parameter estimation
  • Discuss the memory-based neural networks
  • What is the Markov decision process

Research Topics in Human Visual System and Machine Learning

  • The role of video processing experts
  • Understanding the psychology of vision
  • Using the HVS model
  • Discuss the process of Chroma subsampling
  • Image compression techniques
  • The low-pass filter characteristic of the HVS model
  • Describe the human eye
  • The impact of 3D resolution
  • How does a depth-inverted face look like?
  • Brightness resolution
  • Complex visual systems

We have a list of professional writers ready to offer writing assistance in any area of machine learning. Contact us with a “ do my research paper ” request and t ry our cheap but quality service today!  

big data research topics

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Save my name, email, and website in this browser for the next time I comment.

Terms & Conditions Loyalty Program Privacy Policy Money-Back Policy

Copyright © 2013-2024 MyPaperDone.com

  • Bibliography
  • More Referencing guides Blog Automated transliteration Relevant bibliographies by topics
  • Automated transliteration
  • Relevant bibliographies by topics
  • Referencing guides

Dissertations / Theses on the topic 'Machine Learning (ML)'

Create a spot-on reference in apa, mla, chicago, harvard, and other styles.

Select a source type:

  • Journal article
  • Video (online)
  • All types...
  • Archival document
  • Book chapter
  • Complete reference
  • Conference paper
  • Copyright certificate
  • Dictionary entry
  • Dissertation / Thesis
  • Encyclopedia
  • Encyclopedia article
  • Extended abstract of dissertation
  • Newspaper article
  • Press release
  • Religious text
  • Social media post

Consult the top 50 dissertations / theses for your research on the topic 'Machine Learning (ML).'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

Holmberg, Lars. "Human In Command Machine Learning." Licentiate thesis, Malmö universitet, Malmö högskola, Institutionen för datavetenskap och medieteknik (DVMT), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-42576.

Nangalia, V. "ML-EWS - Machine Learning Early Warning System : the application of machine learning to predict in-hospital patient deterioration." Thesis, University College London (University of London), 2017. http://discovery.ucl.ac.uk/1565193/.

John, Meenu Mary. "Design Methods and Processes for ML/DL models." Licentiate thesis, Malmö universitet, Institutionen för datavetenskap och medieteknik (DVMT), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-45026.

Tabell, Johnsson Marco, and Ala Jafar. "Efficiency Comparison Between Curriculum Reinforcement Learning & Reinforcement Learning Using ML-Agents." Thesis, Blekinge Tekniska Högskola, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-20218.

Mattsson, Fredrik, and Anton Gustafsson. "Optimize Ranking System With Machine Learning." Thesis, Högskolan i Halmstad, Akademin för informationsteknologi, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:hh:diva-37431.

Kakadost, Naser, and Charif Ramadan. "Empirisk undersökning av ML strategier vid prediktion av cykelflöden baserad på cykeldata och veckodagar." Thesis, Malmö universitet, Fakulteten för teknik och samhälle (TS), 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-20168.

Sammaritani, Gloria. "Google BigQuery ML. Analisi comparativa di un nuovo framework per il Machine Learning." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2020.

Gustafsson, Sebastian. "Interpretable serious event forecasting using machine learning and SHAP." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-444363.

Schoenfeld, Brandon J. "Metalearning by Exploiting Granular Machine Learning Pipeline Metadata." BYU ScholarsArchive, 2020. https://scholarsarchive.byu.edu/etd/8730.

Hellborg, Per. "Optimering av datamängder med Machine learning : En studie om Machine learning och Internet of Things." Thesis, Högskolan i Skövde, Institutionen för informationsteknologi, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-13747.

Nämerforslund, Tim. "Machine Learning Adversaries in Video Games : Using reinforcement learning in the Unity Engine to create compelling enemy characters." Thesis, Mittuniversitetet, Institutionen för informationssystem och –teknologi, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:miun:diva-42746.

Mahfouz, Tarek Said. "Construction legal support for differing site conditions (DSC) through statistical modeling and machine learning (ML)." [Ames, Iowa : Iowa State University], 2009.

REPETTO, MARCO. "Black-box supervised learning and empirical assessment: new perspectives in credit risk modeling." Doctoral thesis, Università degli Studi di Milano-Bicocca, 2023. https://hdl.handle.net/10281/402366.

Björkberg, David. "Comparison of cumulative reward withone, two and three layered artificialneural network in a simple environmentwhen using ml-agents." Thesis, Blekinge Tekniska Högskola, Institutionen för datavetenskap, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-21188.

Ucci, Graziano. "The Interstellar Medium of Galaxies: a Machine Learning Approach." Doctoral thesis, Scuola Normale Superiore, 2019. http://hdl.handle.net/11384/85928.

Gilmore, Eugene M. "Learning Interpretable Decision Tree Classifiers with Human in the Loop Learning and Parallel Coordinates." Thesis, Griffith University, 2022. http://hdl.handle.net/10072/418633.

Sridhar, Sabarish. "SELECTION OF FEATURES FOR ML BASED COMMANDING OF AUTONOMOUS VEHICLES." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-287450.

Lagerkvist, Love. "Neural Novelty — How Machine Learning Does Interactive Generative Literature." Thesis, Malmö universitet, Fakulteten för kultur och samhälle (KS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-21222.

Bhogi, Keerthana. "Two New Applications of Tensors to Machine Learning for Wireless Communications." Thesis, Virginia Tech, 2021. http://hdl.handle.net/10919/104970.

Garg, Anushka. "Comparing Machine Learning Algorithms and Feature Selection Techniques to Predict Undesired Behavior in Business Processesand Study of Auto ML Frameworks." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-285559.

Hanski, Jari, and Kaan Baris Biçak. "An Evaluation of the Unity Machine Learning Agents Toolkit in Dense and Sparse Reward Video Game Environments." Thesis, Uppsala universitet, Institutionen för speldesign, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-444982.

Stellmar, Justin. "Predicting the Deformation of 3D Printed ABS Plastic Using Machine Learning Regressions." Youngstown State University / OhioLINK, 2020. http://rave.ohiolink.edu/etdc/view?acc_num=ysu1587462911261523.

Krüger, Franz David, and Mohamad Nabeel. "Hyperparameter Tuning Using Genetic Algorithms : A study of genetic algorithms impact and performance for optimization of ML algorithms." Thesis, Mittuniversitetet, Institutionen för informationssystem och –teknologi, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:miun:diva-42404.

Björkman, Desireé. "Machine Learning Evaluation of Natural Language to Computational Thinking : On the possibilities of coding without syntax." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-424269.

Lundin, Lowe. "Artificial Intelligence for Data Center Power Consumption Optimisation." Thesis, Uppsala universitet, Avdelningen för systemteknik, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-447627.

Wessman, Filip. "Advanced Algorithms for Classification and Anomaly Detection on Log File Data : Comparative study of different Machine Learning Approaches." Thesis, Mittuniversitetet, Institutionen för informationssystem och –teknologi, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:miun:diva-43175.

Narmack, Kirilll. "Dynamic Speed Adaptation for Curves using Machine Learning." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-233545.

Mathias, Berggren, and Sonesson Daniel. "Design Optimization in Gas Turbines using Machine Learning : A study performed for Siemens Energy AB." Thesis, Linköpings universitet, Programvara och system, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-173920.

Giuliani, Luca. "Extending the Moving Targets Method for Injecting Constraints in Machine Learning." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2021. http://amslaurea.unibo.it/23885/.

Tarullo, Viviana. "Artificial Neural Networks for classification of EMG data in hand myoelectric control." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2019. http://amslaurea.unibo.it/19195/.

Wallner, Vanja. "Mapping medical expressions to MedDRA using Natural Language Processing." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-426916.

Hellberg, Johan, and Kasper Johansson. "Building Models for Prediction and Forecasting of Service Quality." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-295617.

Nardello, Matteo. "Low-Power Smart Devices for the IoT Revolution." Doctoral thesis, Università degli studi di Trento, 2020. http://hdl.handle.net/11572/274371.

Nardello, Matteo. "Low-Power Smart Devices for the IoT Revolution." Doctoral thesis, Università degli studi di Trento, 2020. http://hdl.handle.net/11572/274371.

Michelini, Mattia. "Barcode detection by neural networks on Android mobile platforms." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2020. http://amslaurea.unibo.it/21080/.

Lundström, Robin. "Machine Learning for Air Flow Characterization : An application of Theory-Guided Data Science for Air Fow characterization in an Industrial Foundry." Thesis, Karlstads universitet, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:kau:diva-72782.

Daneshvar, Saman. "User Modeling in Social Media: Gender and Age Detection." Thesis, Université d'Ottawa / University of Ottawa, 2019. http://hdl.handle.net/10393/39535.

Rosell, Felicia. "Tracking a ball during bounce and roll using recurrent neural networks." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-239733.

Hallberg, Jesper. "Searching for the charged Higgs boson in the tau nu analysis using Boosted Decision Trees." Thesis, Uppsala universitet, Högenergifysik, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-301351.

Klingvall, Emelie. "Artificiell intelligens som ett beslutsstöd inom mammografi : En kvalitativ studie om radiologers perspektiv på icke-tekniska utmaningar." Thesis, Högskolan i Skövde, Institutionen för informationsteknologi, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-18768.

Alatram, Ala'a A. M. "A forensic framework for detecting denial-of-service attacks in IoT networks using the MQTT protocol." Thesis, Edith Cowan University, Research Online, Perth, Western Australia, 2022. https://ro.ecu.edu.au/theses/2561.

Forssell, Melker, and Gustav Janér. "Product Matching Using Image Similarity." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-413481.

Bengtsson, Theodor, and Jonas Hägerlöf. "Stora mängder användardata för produktutveckling : Möjligheter och utmaningar vid integrering av stora mängder användardata i produktutvecklingsprocesser." Thesis, KTH, Integrerad produktutveckling, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-297966.

Nordqvist, My. "Classify part of day and snow on the load of timber stacks : A comparative study between partitional clustering and competitive learning." Thesis, Mittuniversitetet, Institutionen för informationssystem och –teknologi, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:miun:diva-42238.

Mele, Matteo. "Convolutional Neural Networks for the Classification of Olive Oil Geographical Origin." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2020.

Hjerpe, Adam. "Computing Random Forests Variable Importance Measures (VIM) on Mixed Numerical and Categorical Data." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-185496.

Ahlm, Kristoffer. "IDENTIFIKATION AV RISKINDIKATORER I FINANSIELL INFORMATION MED HJÄLP AV AI/ML : Ökade möjligheter för myndigheter att förebygga ekonomisk brottslighet." Thesis, Umeå universitet, Institutionen för matematik och matematisk statistik, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-184818.

Zanghieri, Marcello. "sEMG-based hand gesture recognition with deep learning." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2019. http://amslaurea.unibo.it/18112/.

Benatti, Mattia. "Progettazione e Sviluppo di una Piattaforma Multi-Sorgente per l’Ottimizzazione dei Servizi di Emergenza." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2021.

Sibelius, Parmbäck Sebastian. "HMMs and LSTMs for On-line Gesture Recognition on the Stylaero Board : Evaluating and Comparing Two Methods." Thesis, Linköpings universitet, Artificiell intelligens och integrerade datorsystem, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-162237.

  • Thursday, February 22, 2024

Arabic

  • Artificial Intelligence
  • Deep Learning

Federated Learning

  • IT & Management
  • Cyber Security & Forensics
  • Research Insights
  • 2022 Volume 1
  • 2022 Volume 2
  • 2022 Volume 3
  • 2022 Volume 4
  • 2023 Volume 1
  • 2023 Volume 2
  • 2023 Volume 3
  • 2023 Volume 4
  • 2023 Volume 1, Issue 01
  • 2023 Volume 1, Issue 02
  • 2023 Volume 1, Issue 03
  • 2023 Volume 1, Issue 04
  • 2024 Volume 1 Issue 01
  • 2024 Volume 1 Issue 02
  • 2024 Volume 1 Issue 03
  • 2024 Volume 1 Issue 04
  • 2023 Volume 01 Issue 01
  • 2023 Volume 01 Issue 02
  • Top Computer Science Universities in the World
  • Shanghai University Ranking 2023
  • QS World University Rankings
  • Times World University Ranking
  • Latest NEW Jobs Finder
  • ABDC Ranking
  • FT50 journal
  • ABS Ranking
  • Tech NEW Stories
  • Announcements
  • Upcoming Events
  • Featured Articles
  • Publish NEW with Us
  • Artificial intelligence
  • Machine Learning

Machine Learning Research Topics 2023

By : Aksaht Gaurav , Ronin Institute, U.S

Machine learning (ML) is a rapidly evolving field, with new technologies and approaches being developed at a breakneck pace [1-6]. As we approach the year 2023, the field is poised to make significant advancements in a number of areas. This blog post will explore some of the top machine-learning research topics for 2023.

Explainable Artificial Intelligence (XAI)

XAI is an area of research that focuses on developing machine learning algorithms that can provide clear explanations for their decisions [7-11]. As machine learning becomes more widespread, there is a growing need for algorithms that can be easily understood and interpreted by humans. XAI research is expected to make significant strides in 2023 and beyond, with new models and algorithms that offer more transparency and accountability.

Federated learning is a technique that allows multiple devices to contribute to a shared machine learning model without sending their data to a centralized server [12-16]. This technique has many potential applications, from improving personalized recommendations to developing better predictive models for medical research. In 2023, we can expect to see significant advancements in federated learning, with new algorithms and techniques that improve its efficiency and accuracy.

Continual Learning

Continual learning is an area of research that focuses on developing machine learning algorithms that can learn new tasks without forgetting their previous knowledge [4]. This is an important area of research, as current machine learning models are often unable to learn new tasks without significant retraining. In 2023, we can expect to see significant advancements in continual learning, with new models and algorithms that can learn new tasks more efficiently.

Reinforcement Learning

Reinforcement learning is a type of machine learning that focuses on training agents to make decisions in complex environments [16-20]. In 2023, we can expect to see significant advancements in reinforcement learning, with new techniques that improve the stability and efficiency of training algorithms.

Machine Learning for Healthcare

Machine learning has the potential to revolutionize the healthcare industry, from improving diagnosis and treatment to developing better predictive models for disease outbreaks [6]. In 2023, we can expect to see significant advancements in machine learning for healthcare, with new models and algorithms that offer better accuracy and efficiency.

Machine Learning for Metaverse

The metaverse is an emerging concept that refers to a virtual world that combines elements of gaming, social media, and other online experiences [7]. As the metaverse becomes more widespread, there will be a growing need for machine learning algorithms that can analyze and interpret the vast amounts of data generated by these virtual worlds. In 2023, we can expect to see significant advancements in machine learning for the metaverse, with new models and algorithms that offer better insights and predictions.

Natural Language Processing

Natural language processing (NLP) is an area of machine learning that focuses on developing algorithms that can analyze and understand human language [8]. In 2023, we can expect significant NLP advancements, with new models and algorithms that offer better accuracy and efficiency. This could have many potential applications, from improving voice assistants and chatbots to developing better text-to-speech and speech-to-text systems.

Autonomous Systems

Autonomous systems, such as self-driving cars and drones, rely on machine learning algorithms to make decisions and navigate complex environments [9]. In 2023, we can expect significant advancements in autonomous systems, with new models and algorithms offering better accuracy and safety. This could have many potential applications, from improving transportation and logistics to developing better surveillance and security systems.

Quantum Machine Learning

Quantum computing is an emerging technology that has the potential to revolutionize machine learning. In 2023, we can expect to see significant advancements in quantum machine learning, with new algorithms and techniques that offer better efficiency and scalability[10]. This could have many potential applications, from improving drug discovery and materials science to developing better machine learning models for financial and insurance industries.

Overall, the field of machine learning is rapidly evolving, with many exciting research topics on the horizon. From machine learning for metaverse and autonomous systems to natural language processing and quantum machine learning, there is a lot to look forward to in 2023 and beyond. By staying up to date with the latest research, we can help shape the future of machine learning and its impact on our world.

  • Mitchell, T. M. (2007).  Machine learning  (Vol. 1). New York: McGraw-hill.
  • Sra, S., Nowozin, S., & Wright, S. J. (Eds.). (2012).  Optimization for machine learning . Mit Press.
  • Wang, J., et al., (2022). Pcnncec: Efficient and privacy-preserving convolutional neural network inference based on cloud-edge-client collaboration. IEEE Transactions on Network Science and Engineering.
  • Jordan, M. I., & Mitchell, T. M. (2015). Machine learning: Trends, perspectives, and prospects .  Science ,  349 (6245), 255-260.
  • Gopal Mengi ;  Sudhakar Kumar  (2022)   Artificial Intelligence and Machine Learning in Healthcare , Insights2Tecinfo, pp. 1
  • Mahesh, B. (2020). Machine learning algorithms-a review.   International Journal of Science and Research (IJSR).[Internet] ,  9 , 381-386.
  • Gupta, B. B., et al., (2021, October). A big data and deep learning based approach for ddos detection in cloud computing environment. In 2021 IEEE 10th Global conference on consumer electronics (GCCE) (pp. 287-290). IEEE.
  • Carleo, G., Cirac, I., Cranmer, K., Daudet, L., Schuld, M., Tishby, N., … & Zdeborová, L. (2019). Machine learning and the physical sciences .  Reviews of Modern Physics ,  91 (4), 045002.
  • Shavlik, J. W., Dietterich, T., & Dietterich, T. G. (Eds.). (1990).   Readings in machine learning . Morgan Kaufmann.
  • Rajput, R. K. S., et al., (2022). Cloud data centre energy utilization estimation: Simulation and modelling with idr .  International Journal of Cloud Applications and Computing (IJCAC) ,  12 (1), 1-16.
  • El Naqa, I., & Murphy, M. J. (2015).  What is machine learning?  (pp. 3-11). Springer International Publishing.
  • Mishra A, (2022)  Analysis of the Development of Big data and AI-Based Technologies for the Cloud Computing Environment , Data Science Insights Magazine, Insights2Techinfo, Volume 2, pp. 9-12. 2022.
  • Pai, M. L., et al., (2020). Application of Artificial Neural Networks and Genetic Algorithm for the Prediction of Forest Fire Danger in Kerala . In  Intelligent Systems Design and Applications: 18th International Conference on Intelligent Systems Design and Applications (ISDA 2018) held in Vellore, India, December 6-8, 2018, Volume 2  (pp. 935-942). Springer International Publishing.
  • Ahamed, J., et al., (2022). CDPS-IoT: cardiovascular disease prediction system based on iot using machine learning.
  • Provost, F., & Kohavi, R. (1998). On applied research in machine learning.   MACHINE LEARNING-BOSTON- ,  30 , 127-132.
  • Mishra, A., et al., (2011, September). A comparative study of distributed denial of service attacks, intrusion tolerance and mitigation techniques. In 2011 European Intelligence and Security Informatics Conference (pp. 286-289). IEEE.
  • Gupta, B. B., et al., (2011). On estimating strength of a DDoS attack using polynomial regression model. In Advances in Computing and Communications: First International Conference, ACC 2011, Kochi, India, July 22-24, 2011, Proceedings, Part IV 1 (pp. 244-249). Springer Berlin Heidelberg.
  • Alpaydin, E. (2016).  Machine learning: the new AI . MIT press.
  • Ayodele, T. O. (2010). Types of machine learning algorithms .  New advances in machine learning ,  3 , 19-48.
  • Akash Sharma et al., (2022) Classical Computer to Quantum Computers , Insights2Tecinfo, pp. 1

A. Gaurav (2023) Machine Learning Research Topics 2023 , Insights2Techinfo, pp.1

Leave a Reply Cancel reply

Your email address will not be published.

Privacy Overview

IMAGES

  1. Thesis topics in machine learning by Techsparks

    machine learning thesis topics

  2. Latest Thesis Topics in Machine Learning for Research Scholars

    machine learning thesis topics

  3. Top 5 Thesis Topics for Machine Learning [Customized Research Support]

    machine learning thesis topics

  4. 10 Machine Learning Project (Thesis) Topics for 2020

    machine learning thesis topics

  5. Hot topics in machine learning for research and thesis by Write

    machine learning thesis topics

  6. Top 10 Research and Thesis Topics for ML Projects in 2022

    machine learning thesis topics

VIDEO

  1. Research Methodology part two

  2. Lec07_part1_introduction to machine learning

  3. How do I write my PhD thesis about Artificial Intelligence, Machine Learning and Robust Clustering?

  4. What Does Machine Learning Engineer Actually Do? Part 2

  5. Empowering Music Creation with Machine Learning (Thesis Proposal)

COMMENTS

  1. Top 10 Research and Thesis Topics for ML Projects in 2022

    Voice and Speech Recognition, Signal Processing, Message Embedding, Message Extraction from Voice Encoded, and more are the best research and thesis topics for ML projects. Sentiment Analysis Sentiment analysis is one of the best Machine Learning projects well-known to uncover emotions in the text.

  2. Available Master's thesis topics in machine learning

    Machine Learning Education thesis topics Available Master's thesis topics in machine learning Here we list topics that are available. You may also be interested in our list of completed Master's theses. Learning and inference with large Bayesian networks Sum-product networks Bayesian Bayesian networks

  3. The Future of AI Research: 20 Thesis Ideas for Undergraduate ...

    This article provides a list of 20 potential thesis ideas for an undergraduate program in machine learning and deep learning in 2023. Each thesis idea includes an introduction, which presents a brief overview of the topic and the research objectives.

  4. Exploring 250+ Machine Learning Research Topics

    Real-world Applications Machine learning research has brought about tangible changes in our daily lives. Voice assistants like Siri and Alexa, recommendation systems on streaming platforms, and personalized healthcare diagnostics are just a few examples of how this research impacts our world.

  5. PhD Dissertations

    Data-driven Decisions - An Anomaly Detection Perspective Shubhranshu Shekhar, 2023. METHODS AND APPLICATIONS OF EXPLAINABLE MACHINE LEARNING Joon Sik Kim, 2023. Applied Mathematics of the Future Kin G. Olivares, 2023. NEURAL REASONING FOR QUESTION ANSWERING Haitian Sun, 2023.

  6. PDF Undergraduate Fundamentals of Machine Learning

    Broadly, machine learning is the application of statistical, mathematical, and numerical techniques to derive some form of knowledge from data. This 'knowledge' may a ord us some sort of summarization, visualization, grouping, or even predictive power over data sets.

  7. 10 Compelling Machine Learning Ph.D. Dissertations for 2020

    1. Bayesian Modeling and Variable Selection for Complex Data As we routinely encounter high-throughput data sets in complex biological and environmental research, developing novel models and methods for variable selection has received widespread attention.

  8. Thesis Topic Proposals 2022-2023

    Below you can see the thesis topics for 2022-2023. We offer 3 different thesis formats: - Format 1 : Regular thesis (fully supervised by KU Leuven) - Format 2 : Thesis in cooperation with a company (supervised by KU Leuven and the company) - Format 3 : Thesis with a company project within a company (supervised by the company)

  9. 17 Compelling Machine Learning Ph.D. Dissertations

    This machine learning dissertation analyzes data to build a quantitative understanding of the world. Linear algebra is the foundation of algorithms, dating back one hundred years, for extracting structure from data.

  10. 37 Research Topics In Data Science To Stay On Top Of » EML

    3.) Auto Machine Learning. Auto machine learning is a research topic in data science concerned with developing algorithms that can automatically learn from data without intervention. This area of research is vital because it allows data scientists to automate the process of writing code for every dataset.

  11. Thesis Topics

    Thesis Topics. This list includes topics for potential bachelor or master theses, guided research, projects, seminars, and other activities. Search with Ctrl+F for desired keywords, e.g. 'machine learning' or others. PLEASE NOTE: If you are interested in any of these topics, click the respective supervisor link to send a message with a ...

  12. 8 Best Topics for Research and Thesis in Artificial Intelligence

    1. Machine Learning Machine Learning involves the use of Artificial Intelligence to enable machines to learn a task from experience without programming them specifically about that task. (In short, Machines learn automatically without human hand holding!!!)

  13. Undergraduate Fundamentals of Machine Learning

    Abstract. Drawing on lectures, course materials, existing textbooks, and other resources, we synthesize and consolidate the content necessary to offer a successful first exposure to machine learning for students with an undergraduate-level background in linear algebra and statistics. The final product is a textbook for Harvard's introductory ...

  14. Information Systems and Machine Learning Lab, University of Hildesheim

    The notion of attention in machine learning approaches proved successful in domain adaptation, time series forecasting, and image classification. However, all attention techniques are designed on an instance level. In this thesis, you will investigate the a new form of attention, by means of dataset meta-feature similarity.

  15. machine learning Latest Research Papers

    Find the latest published documents for machine learning, Related hot topics, top authors, the most cited documents, and related journals

  16. Open Theses

    Our topics lie at the intersection of generative models like diffusion/flow matching models graph representation learning, e.g., graph neural networks.

  17. Latest Thesis Topics in Machine Learning for Research Scholars

    Below is the list of the latest thesis topics in Machine learning for research scholars: The classification technique for the face spoof detection in artificial neural networks using concepts of machine learning. The iris detection and reorganization system using classification and glcm algorithm in machine learning.

  18. 10 Machine Learning Project (Thesis) Topics for 2020

    Explore 10 different topics for your machine learning project, thesis or dissertation, such as classification, clustering, regression and sentiment analysis. Each topic has a link to the dataset and some support from the author.

  19. Thesis Topics for Machine Learning

    Let us now have an idea about various headings to be included in any thesis topics for machine learning. Introduction - overview of the thesis; Related / Existing works - presents of existing research; Problems definition/statements - identify and highlight the problems; Research methodology - convey the proposed concepts; Results and Discussion - discuss the results of the proposed ...

  20. Intelligent Autonomous Systems

    Learning Latent Representations for Embodied Agents. Scope: Master Thesis Advisor: Michael Drolet, Oleg Arenz Added: 2023-11-05 Start: ASAP Topic: Learning from Demonstration [1] and Policy Search [2] are fundamental approaches in training robot policies, but there's a critical challenge: choosing the right expert. Typically, experts are expected to operate within the same state space and ...

  21. 147 Outstanding Machine Learning Topics For Students

    147 Top Machine Learning Topics To Get Your Paper Easy First of all, let's determine - what is machine learning? Generally, it refers to the study of computer algorithms that improve automatically through the use of data and experience. Machine learning is seen as part of AI that makes decisions or predictions without being entirely programmed.

  22. Dissertations / Theses: 'Machine Learning (ML)'

    This thesis proposes a mode of inquiry that considers the inter- active qualities of what machine learning does, as opposed the tech- nical specifications of what machine learning is. A shift in focus from the technicality of ML to the artifacts it creates allows the interaction designer to situate its existing skill set, affording it to engage ...

  23. Machine Learning Research Topics 2023

    Natural Language Processing Natural language processing (NLP) is an area of machine learning that focuses on developing algorithms that can analyze and understand human language [8]. In 2023, we can expect significant NLP advancements, with new models and algorithms that offer better accuracy and efficiency.