Proposal Exam: Yaesop Lee

Friday, December 9, 2022
3:00 p.m.
AVW 2224
Maria Hoo
301 405 3681
mch@umd.edu

ANNOUNCEMENT: Ph.D. Research Proposal Exam

 

Name: Yaesop Lee

 

Committee:

Professor Shuvra S. Bhattacharyya (Chair)

Professor Manoj Franklin

Professor Jonathan Z. Simon

Date/time : 12/09/2022 3PM

 

Location : AVW 2224

 

Title: Design Techniques for Embedded Computer Vision and Signal Processing

 

Abstract: In this proposal, we explore new design techniques to facilitate the implementation

of efficient deep learning systems for embedded computer vision and signal
processing. The techniques are developed to address concerns of real-time
processing efficiency and energy efficiency under resource-constrained operation as
well accuracy considerations, which are conventionally associated with the
development of deep learning solutions. We study two specific application areas for
efficient deep learning --- (1) neural decoding and (2) object detection from multi-
view images, such as those acquired from unmanned aerial vehicles (UAVs).

To address the challenges of efficient deep learning systems, we apply dataflow
based methods for design and implementation of signal and information processing
systems. Signal-processing oriented dataflow concepts provide an efficient
computational model that allows flexibility and expandability to facilitate design and
implementation of complex signal and information processing systems.
In dataflow modeling, applications are modeled as directed graphs, called dataflow
graphs, in which vertices (actors) correspond to discrete computations that are
executed and edges represent communication between pairs of actors. In the first
part of the proposal, we study in depth a recently-introduced model of computation,
called passive-active flow graphs (PAFGs), which can be used in conjunction with
dataflow modeling to facilitate more efficient implementation of dataflow graphs.
In the second part of the proposal, we present the application of dataflow techniques
to develop a novel system for real-time neural decoding. Neural decoding involves
processing signals acquired from the brain --- for example, through calcium imaging
technology --- to predict behavior variables. We refer to the developed system as the
Neuron Detection and Signal Extraction Platform (NDSEP). NDSEP incorporates
streamlined subsystems for neural decoding that are integrated efficiently through
dataflow modeling. The dataflow-based software architecture of NDSEP provides
modularity and extensibility to experiment with alternative modules for neural signal
processing. Our system design also facilitates optimization of trade-offs between
accuracy and real-time performance.

In the third part of the proposal, we address the problem of limited training data,
which is a significant problem for many application areas of embedded computer
vision, especially areas that are highly specialized or are at the very forefront of
computer vision technology. We address this problem specifically in the context of
deep learning for object detection from multi-view images acquired from unmanned
aerial vehicles (UAVs). To help overcome the shortage of relevant training data in
this class of object detection scenarios, we introduce a new dataset and associated
metadata, which integrates real and synthetic data to provide a much larger
collection of labeled data than what is available from real data alone.
We also apply the developed dataset to conduct comprehensive studies of how the
critical attributes of UAV-based images affect machine learning models, and how
these insights can be applied to advance the training and testing of the models.
Moreover, in the fourth part of the proposal, we explore fundamental algorithm
development for efficient object detection from multi-view images. In this work,
we propose a simplified 2-dimensional object detection technique that can be
implemented to leverage multiple images for a scene. This work provides a simple
but effective way to extend the detection architecture for a single view image to an
architecture for multi-view images. A useful feature of the proposed approach is
that it requires only a minimal amount of additional computation to extend an
architecture from single- to multi-view operation.

Our proposed future work involves two main directions. First, building on our
completed work on neural signal processing, we propose to investigate efficient
incremental learning for deep-learning-based neural decoding. Incremental deep
learning is important for practical neural decoding systems to enable adaptation of
the systems to operational environments that have different characteristics from the
corresponding training environments. We also plan to explore methods for dataflow-
based design and implementation of adaptive, real-time systems for object detection
and tracking. In a distributed object detection environment, the proposed system will
adaptively control knowledge extraction to provide acceptable object detection
accuracy while minimizing energy consumption across the distributed network. The
underlying dataflow framework will model an adaptive embedded computer vision
system in terms of multiple subgraphs, where subgraphs are carefully selected for
execution and scheduled dynamically to optimize performance under the given
resource constraints.

Audience: Faculty 

remind we with google calendar

 

April 2024

SU MO TU WE TH FR SA
31 1 2 3 4 5 6
7 8 9 10 11 12 13
14 15 16 17 18 19 20
21 22 23 24 25 26 27
28 29 30 1 2 3 4
Submit an Event