Calendar

Apr
29
Thu
Thesis Proposal: Michelle Graham
Apr 29 @ 3:00 pm
Thesis Proposal: Michelle Graham

Note: This is a virtual presentation. Here is the link for where the presentation will be taking place.

Title: Photoacoustic imaging to detect major blood vessels and nerves during neurosurgery and head and neck surgery

Abstract: Real-time intraoperative guidance during minimally invasive neurosurgical and head and neck procedures is often limited to endoscopy, CT-guided image navigation, and electromyography, which are generally insufficient to locate major blood vessels and nerves hidden by tissue. Accidental damage to these hidden structures has incidence rates of 6.8% in surgeries to remove pituitary tumors (i.e., endonasal transsphenoidal surgery) and 3-4% in surgeries to remove parotid tumors (i.e., parotidectomy), often resulting in severe consequences, such as patient blindness, paralysis, and death. Photoacoustic imaging is a promising emerging imaging technique to provide real-time guidance of subsurface blood vessels and nerves during these surgeries.

Limited optical penetration through bone and the presence of acoustic clutter, reverberations, aberration, and attenuation can degrade photoacoustic image quality and potentially corrupt the usefulness of this promising intraoperative guidance technique. In order to mitigate image degradation, photoacoustic imaging system parameters may be adjusted and optimized to cater to the specific imaging environment. In particular, parameter adjustment can be categorized into the optimization of photoacoustic signal generation and the optimization of photoacoustic image formation (i.e., beamforming) and image display methods.

In this talk, I will describe my contributions to leverage amplitude- and coherence-based beamforming techniques to improve photoacoustic image display for the detection of blood vessels during endonasal transsphenoidal surgery. I will then present my contributions to the derivation of a novel photoacoustic spatial coherence theory, which provides a fundamental understanding critical to the optimization of coherence-based photoacoustic images. Finally, I will present a plan to translate this work from the visualization of blood vessels during neurosurgery to the visualization of nerves during head and neck surgery. Successful completion of this work will lay the foundation necessary to introduce novel, intraoperative, photoacoustic image guidance techniques that will eliminate the incidence of accidental injury to major blood vessels and nerves during minimally invasive surgeries.

Committee Members:

  • Muyinatu Bell, Department of Electrical and Computer Engineering
  • Xindge Li, Department of Biomedical Engineering
  • Jin Kang, Department of Electrical and Computer Engineering
May
10
Mon
Dissertation Defense: Jordi Abante
May 10 @ 3:00 pm
Dissertation Defense: Jordi Abante

Note: This is a virtual presentation. Here is the link for where the presentation will be taking place.

Title: Statistical Signal Processing Methods for Epigenetic Landscape Analysis

Abstract: Since the DNA structure was discovered in 1953, a great deal of effort has been put into studying this molecule in detail. We now know DNA comprises an organism’s genetic makeup and constitutes a blueprint for life. The study of DNA has dramatically increased our knowledge about cell function and evolution and has led to remarkable discoveries in biology and medicine.

Just as DNA is replicated during cell division, several chemical marks are also passed onto progeny during this process. Epigenetics studies these marks and represents a fascinating research area given their crucial role. Among all known epigenetic marks, 5mc DNA methylation is probably one of the most important ones given its well-established association with various biological processes, such as development and aging, and disease, such as cancer. The work in this dissertation focuses primarily on this epigenetic mark, although it has the potential to be applied to other heritable marks.

In the 1940s, Waddington introduced the term epigenetic landscape to conceptually describe cell pluripotency and differentiation. This concept lived in the abstract plane until Jenkinson et al. 2017 & 2018 estimated actual epigenetic landscapes from WGBS data, and the work led to startling results with biological implications in development and disease. Here, we introduce an array of novel computational methods that draw from that work. First, we present CPEL, a method that uses a variant of the original landscape proposed by Jenkinson et al., which, together with a new hypothesis testing framework, allows for the detection of DNA methylation imbalances between homologous chromosomes. Then, we present CpelTdm, a method that builds upon CPEL to perform differential methylation analysis between groups of samples using targeted bisulfite sequencing data. Finally, we extend the original probabilistic model proposed by Jenkinson et al. to estimate methylation landscapes and perform differential analysis from nanopore data.

Overall, this work addresses immediate needs in the study of DNA methylation. The methods presented here can lead to a better characterization of this critical epigenetic mark and enable biological discoveries with implications for diagnosing and treating complex human diseases.

Committee Members

  • John Goutsias, Department of Electrical and Computer Engineering
  • Archana Venkataraman, Department of Electrical and Computer Engineering
  • Sanjeev Khudanpur, Department of Electrical and Computer Engineering
May
24
Mon
Dissertation Defense: Xing Di
May 24 @ 12:00 pm
Dissertation Defense: Xing Di

Note: This is a virtual presentation. Here is the link for where the presentation will be taking place.

Title: Deep Learning Based Face Image Synthesis

Abstract: Face image synthesis is an important problem in the biometrics and computer vision communities due to its applications in law enforcement and entertainment. In this thesis, we develop novel deep neural network models and associated loss functions for two face image synthesis problems, namely thermal to visible face synthesis and visual attribute to face synthesis.

In particular, for thermal to visible face synthesis, we propose a model which makes use of facial attributes to obtain better synthesis. We use attributes extracted from visible images to synthesize attribute-preserved visible images from thermal imagery. A pre-trained attribute predictor network is used to extract attributes from the visible image. Then, a novel multi-scale generator is proposed to synthesize the visible image from the thermal image guided by the extracted attributes. Finally, a pre-trained VGG-Face network is leveraged to extract features from the synthesized image and the input visible image for verification.

In addition, we propose another thermal to visible face synthesis method based on a self-attention generative adversarial network (SAGAN) which allows efficient attention-guided image synthesis. Rather than focusing only on synthesizing visible faces from thermal faces, we also propose to synthesize thermal faces from visible faces. Our intuition is based on the fact that thermal images also contain some discriminative information about the person for verification. Deep features from a pre-trained Convolutional Neural Network (CNN) are extracted from the original as well as the synthesized images. These features are then fused to generate a template which is then used for cross-modal face verification.

Regarding attribute to face image synthesis, we propose the Att2SK2Face model for face image synthesis from visual attributes via sketch. In this approach, we first synthesize a facial sketch corresponding to the visual attributes and then generate the face image based on the synthesized sketch. The proposed framework is based on a combination of two different Generative Adversarial Networks (GANs) – (1) a sketch generator network which synthesizes realistic sketch from the input attributes, and (2) a face generator network which synthesizes facial images from the synthesized sketch images with the help of facial attributes.

Finally, we propose another synthesis model, called Att2MFace, which can simultaneously synthesize multimodal faces from visual attributes without requiring paired data in different domains for training the network. We introduce a novel generator with multimodal stretch-out modules to simultaneously synthesize multimodal face images. Additionally, multimodal stretch-in modules are introduced in the discriminator which discriminates between real and fake images.

Committee Members

  • Vishal Patel, Department of Electrical and Computer Engineering
  • Rama Chellappa, Department of Electrical and Computer Engineering
  • Carlos Castillo, Department of Electrical and Computer Engineering
May
25
Tue
Dissertation Defense: Arun Nair
May 25 @ 12:30 pm
Dissertation Defense: Arun Nair

Note: This is a virtual presentation. Here is the link for where the presentation will be taking place.

Title: Machine Learning for Beamforming in Ultrasound, Radar, and Audio

Abstract: Multi-sensor signal processing plays a crucial role in the working of several everyday technologies, from correctly understanding speech on smart home devices to ensuring aircraft fly safely. A specific type of multi-sensor signal processing called beamforming forms a central part of this thesis. Beamforming works by combining the information from several spatially distributed sensors to directionally filter information, boosting the signal from a certain direction but suppressing others. The idea of beamforming is key to the domains of ultrasound, radar, and audio.

Machine learning, succinctly defined by Tom Mitchell as “the study of algorithms that improve automatically through experience” is the other central part of this thesis. Machine learning, especially its sub-field of deep learning, has enabled breakneck progress in tackling several problems that were previously thought intractable. Today, machine learning powers many of the cutting edge systems we see on the internet for image classification, speech recognition, language translation, and more.

In this dissertation, we look at beamforming pipelines in ultrasound, radar, and audio from a machine learning lens and endeavor to improve different parts of the pipelines using ideas from machine learning. Starting off in the ultrasound domain, we use deep learning as an alternative to beamforming in ultrasound and improve the information extraction pipeline by simultaneously generating both a segmentation map and B-mode image of high quality directly from raw received ultrasound data.

Next, we move to the radar domain and study how deep learning can be used to improve signal quality in ultra-wideband synthetic aperture radar by suppressing radio frequency interference, random spectral gaps, and contiguous block spectral gaps. By training and applying the networks on raw single-aperture data prior to beamforming, it can work with myriad sensor geometries and different beamforming equations, a crucial requirement in synthetic aperture radar.

Finally, we move to the audio domain and derive a machine learning inspired beamformer to tackle the problem of ensuring the audio captured by a camera matches its visual content, a problem we term audiovisual zoom. Unlike prior work which is capable of only enhancing a few individual directions, our method enhances audio from a contiguous field of view.

Committee Members

  • Trac Tran, Department of Electrical and Computer Engineering
  • Muyinatu Bell, Department of Electrical and Computer Engineering
  • Vishal Patel, Department of Electrical and Computer Engineering
May
28
Fri
Dissertation Defense: Takeshi Uejima
May 28 @ 9:00 am
Dissertation Defense: Takeshi Uejima

Note: This is a virtual presentation. Here is the link for where the presentation will be taking place.

Title: A Unified Visual Saliency Model for Neuromorphic Implementation

Abstract: Although computer capabilities have expanded tremendously, a significant wall remains between the computer and the human brain. The brain can process massive amounts of information obtained from a complex environment and control the entire body in real time with low energy consumption. This thesis tackles this mystery by modeling and emulating how the brain processes information based on the available knowledge of biological and artificial intelligence as studied in neuroscience, cognitive science, computer science, and computer engineering.

Saliency modeling relates to visual sense and biological intelligence. The retina captures and sends much data about the environment to the brain. However, as the visual cortex cannot process all the information in detail at once, the early stages of visual processing discard unimportant information. Because only the fovea has high-resolution imaging, individuals move their eyeballs in the direction of the important part of the scene. Therefore, eye movements can be thought of as an observable output of the early visual process in the brain. Saliency modeling aims to understand this mechanism and predict eye fixations.

Researchers have built biologically plausible saliency models that emulate the biological process from the retina through the visual cortex. Although many saliency models have been proposed, most are not bio-realistic. This thesis models the biological mechanisms for the perception of texture, depth, and motion. While texture plays a vital role in the perception process, defining texture in a mathematical way is not easy. Thus, it is necessary to build an architecture of texture processing based on the biological perception mechanism. Binocular stereopsis is another intriguing function of the brain. While scholars have evaluated many computational algorithms for stereovision, pursuing biological plausibility means implementing a neuromorphic method into a saliency model. Motion is another critical clue that helps animals survive. In this thesis, the motion feature is implemented in a bio-realistic way based on neurophysiological observation.

Moreover, the thesis will integrate these processes and propose a unified saliency model that can handle 3D dynamic scenes in a similar way to how the brain deals with the real environment. Thus, this investigation will use saliency modeling to examine intriguing properties of human visual processing and discuss how the brain achieves this remarkable capability.

Committee Members

  • Ralph Etienne-Cummings, Department of Electrical and Computer Engineering
  • Andreas Andreou, Department of Electrical and Computer Engineering
  • Philippe Pouliquen, Department of Electrical and Computer Engineering
  • Ernst Niebur, Department of Neuroscience
Back to top