Syed Imran Ahmed
Syed Imran Ahmed
Abstract
Human motion acknowledgment is a well-known examination subject in PC imaginative and
prescient that has until now been extensively considered. Be that as it may, it is as but a
functioning lookup subject for the reason that it assumes a huge phase in several modern and
springing up proper world wise frameworks, as visible commentary and human pc
connection. The exercise awareness hassle has lately been addressed the use of Deep
Reinforcement Learning (DRL) for a range of purposes, consisting of finding interest in
video information or identifying the ideal community structure. DRL-based human
undertaking attention is a new and difficult vicinity of lookup that has solely been round for a
quick time. Along these lines, to work with extra examination in this field, we have
developed a entire overview on motion acknowledgment techniques that include profound
assist learning. At the conclusion of this survey, we supply a precis of the most sizeable
limitations and unsolved troubles in this area that researchers may additionally desire to
tackle in the future.
1. Introduction
The difficult place of lookup recognized as human pastime attention (HAR) focuses on
figuring out and recognizing human things to do or movements. Despite the reality that
humans can without problems pick out the things to do they see in videos, thoroughly
automating this method is tough however indispensable for a extensive vary of real-world
purposes such as video surveillance, interaction between humans and robots, health
monitoring, sports activity analysis, and monitoring of older people's endeavours. A range of
simply on hand sensors, along with RGB and depth cameras are successful of taking pictures
terrific data. to wearable sensors, for example, accelerometers and spinners.
As a result, HAR can take many forms, including: The most extensively used modality, RGB
data, skeleton data, which incorporates two or three dimensions of the places of physique
joints, radar, and WIFI. Human pastime consciousness is formulated as a getting to know
trouble in the pc imaginative and prescient literature, and desktop getting to know methods
are used to classify and apprehend the activities. Researchers had been chiefly centered on
creating handmade features, such as pastime factors and action descriptors, that are successful
of taking pictures the records in movies prior to the emergence of Deep Learning (DL)
fashions in exercise cognizance methods Since setting apart hand created highlights is work
escalated and requires house information, DL techniques commenced being acquainted with
the area for keeping apart fine video portrayals and as a result increasing generalization. This
led to enormous developments in the field.
Enlivened via the manner in which human beings discern out how to act ideally in a number
conditions, guide gaining knowledge of calculations discern out how to play out a unique task
and accomplish a complicated goal thru cooperation with a climate. An agent that learns a
coverage by using exploring the surroundings receives a reward sign that is in line with the
final objective(s) of the agent in each RL algorithm. The specialist's anticipation is boosting
what is known as the Worth, a daily prize. An acronym for a Markov Decision Process
(MDP) is a triple (S, A, R), where S stands for the environment's state, A for the collection of
all possible actions 2, and R for the reward. This is how RL issues are often modelled. In one
RL episode step, the agent completes a task, obtains a reward, and modifies states. Then, the
incentive is utilised to control the coverage so that, ultimately, the agent exhibits its excellent
behaviour (the policy).
A team of RL methods acknowledged as deep reinforcement getting to know deal with high-
dimensional action/state areas by using utilising effective feature approximators recognized
as deep neural networks has proven that deep RL has been very successful. There are two
essential classifications of DRL procedures: coverage gradient and value-function-based
methods, respectively. The first class of strategies tries to estimate every state's fee and locate
the most country fee feature to pick the exceptional coverage π*. The most excellent kingdom
fee feature V*(s), the place s is the contemporary country of the environment, is the most
suitable policy. One of the most widespread processes in this team is Deep QNetwork
(DQN), which combines ordinary Q-learning with deep learning. In some Atari games, DQN
has done superhuman results. Policy gradient methods, on the different hand, parameterize
the coverage and try to immediately optimize it in relation to the anticipated reward.
The two widely used algorithms for this categorization are the Entertainer Pundit and the
Monte Carlo method inclination (Support). In contrast to REINFORCE, Actor-Critic uses
bootstrapping to obtain the cumulative reward rather than waiting until the end of each
episode. Additionally, though they have a tendency to converge to the nearby optimum,
coverage gradient strategies normally converge extra quickly. Policy gradient techniques can
mannequin actions' probabilities, however DQN can't examine stochastic policies. Last
however no longer least, coverage gradient techniques can mannequin a non-stop motion
space, whereas DQN requires a highly-priced motion discretization process.
3. Deep Learning
This interplay is massive but extraordinarily tedious and difficult to do. It would be higher to
have calculations that work with the issue. Profound gaining knowledge of techniques are one
tremendous solutions for control excessive factor statistics and pay attention discriminative
records from the information. The system of extracting statistics representations, or function
extraction, can be carried out by means of deep getting to know algorithms automatically.
The illustration is bought when facts are immediately fed into deep nets except the want for
human intervention (also acknowledged as computerized function extraction).
This crucial section of profound mastering designs is directed to boost closer to these
calculations
that are the goal Man-made reasoning (artificial intelligence), downplaying the world round
free of grasp records and impedance. In outline, profound getting to know endeavours to
show plain degree reflections in facts utilising profound corporations of administered and
moreover unaided gaining knowledge of calculations, to reap from quite a number ranges of
reflections. For a range of purposes, together with classification, it learns deep architectures
with hierarchical representations.
Profound studying fashions include several layers of portrayals. Autoencoders, Restricted
Boltzmann Machines (RBMs), and convolutional layers are just a few of its components. The
uncooked statistics are fed into a community with more than one layers at some stage in
training. The deep network's subsequent layers take as inputs the nonlinear function
transformations that every layer produces. Constricting classifiers or purposes that gain from
summary facts illustration in a hierarchical manner as inputs can make use of the last layer's
output illustration for extended efficiency and performance. Each layer tries to research and
extract underlying explanatory elements by way of making use of a nonlinear transformation
to its input. As a result, a hierarchy of summary representations is realized with the aid of this
process. For instance, when a deep leaning algorithm is used in functions for photo
processing, the image's pixels serve as the first layer, permitting customers to analyse the
edges of a range of objects in the image. Complex aspects like object components
(combination of edges) are represented with the aid of the first layer in the 2nd layer. In order
to assemble object models, the 0.33 layer assembles object parts, which are greater
complicated features.
The profound autoencoders are pretty perchance of the most eminent work in solo profound
aspect learning. They strive to study a illustration of the unique records and are a kind of
synthetic neural network. In order to reconstruct the enter from the encoding, an autoencoder,
auto associator, or Diabolo community is skilled to research high quality encoding.
In reality, the goal end result of the corporation is something comparable of the information.
There are commonly three layers: enter layer, for characteristic vector input; secret layer,
utilized for addressing deliberate elements, and end result layer which use to tackle
reproduced input. The stochastic gradient descent technique, a typical backpropagation
method, is used to analyse parameters for various neural networks. If the autoencoder's
structure only includes one linear hidden layer and the suggest squared error criterion is used
as the loss characteristic to instruct the community, it will function similarly to the
fundamental aspects evaluation (PCA) approach and study the first ok precept factors of the
records.
Convolutional Brain Organizations (CNNs) are organized in the classification of the managed
profound factor gaining knowledge of models. LeCun et al.'s lookup [may have been one of
the earliest on CNNs]. They utilized CNNs to pick out written by means of hand characters.
They had been capable to use CNNs in different purposes like object cognizance and
detection in photograph and speech recognition, as properly as time collection thanks to
developments in computing power.
Convolutional networks have many layers, and their connections intention to examine how to
characterize facets in a hierarchical way. The three main methods used to avoid distortion and
translation to some extent are local receptive fields, shared weights, and spatial or temporal
sub-sampling enter characteristic vectors.
Figure.1 Architecture of a typical convolutional neural network
The CNNs' essential structure is depicted in Figure 1. The subsampling and convolutional
layers are the CNNs' first two layers. To produce characteristic maps, the convolutional layer
makes use of convolution. In order to reap distortion invariance, it makes use of
Neighbourhood receptive fields (small filter sizes) and shared weights (filter maps of same
size). After then, the convolution activity's outcome is long passed through a nonlinear
initiation functionality that is seen through a subsampling layer. The subsampling layer
performs local averaging, also known as max-pooling, which lowers the dimensionality of the
succeeding characteristic maps while maintaining distortion invariance.
The sequence of convolutional and subsampling layers can be used simultaneously in CNN
architectures created for a given application. The output of the last subsampling layer is
supplied to an absolutely connected layer for classification or attention tasks. The newspaper
is suggested for curious readers looking for more information on CNNs. However, in order
for CNNs to be applicable in real applications, particularly in those requiring applications
with highly layered input information, such as handling images and conversations, and to
achieve lifelike performance in examination with shallow pupil techniques, these types of
sophisticated businesses must be given the extensive amount of information. High-
performance computing energy is additionally required due to the fact the deep structure has
so many parameters to train.
Due to the aforementioned requirements, they had now not been extensively used up till
recently. With the creation of notably parallel Graphical Processing Units (GPUs) and the
availability of increasingly more giant statistics sets, these troubles can now be effortlessly
resolved. With the ImageNet database a large-scale picture database with tens of millions of
labelled high-resolution pix divided into lots of categories, and parallel GPUs, deep CNN
architectures may additionally outperform modern sample attention algorithms, for example,
in the area of photograph and imaginative and prescient lookup.
Another profound regulated consist of studying (as properly as unaided) calculation which is
utilized for successive data, the place enter records are dependent on one another in the
manner in which they are developing (information stream) or have learned (words in a
phrase), is repetitive Genius companies (RNNs). Recurrent neural networks (RNNs), in
distinction to feedforward neural networks (FNNs), can have interior states thanks to remarks
connections. This implies that they have a reminiscence which can hold facts about previous
data sources, empowering them to be beneficial for these applications, for example, discourse
acknowledgment which has fleeting and consecutive information.
Methods for DRL-based human undertaking awareness are mentioned in this section. In all
methodologies, there is an mission that can be viewed as a hunt difficulty with no ground-
reality of the perfect arrangement. As a result, the answer ought to be located by way of
interacting with the environment, and reinforcement getting to know and DRL techniques can
flawlessly mannequin the hassle and remedy it. The synopsis of methods is introduced in
Table 1.
Two standards are used to select a set range of frames: the chosen frames' ability for
discrimination as nicely as their connection to the recreation sequence as a whole. To assume
about to understand the representation, a layout-based convolutional intelligence community
is used to create a sketch and a binary mask denoting the selected frames. An agent with a
totally related layer receives the state, and the moves are described via a SoftMax function.
There are three manageable things to do for every part determination: cross to the proper (for
instance choose the following casing on the right), cross to the left (for instance pick out the
following casing on the left), and stay (for instance hold the as of now chosen outline). The
agent receives a reward decided by way of a educated baseline classifier after performing the
action. If the output category label adjustments from mistaken to correct, there is a full-size
high quality reward. If the result shifts from right to incorrect, the agent receives a big poor
reward (punishment). The agent receives a reward of +1 or -1, relying on whether or not the
likelihood of choosing the right category will increase or decreases, if the classification label
stays unchanged. A approach inclination calculation is utilized to put together the RL
specialist.
Frame resolution is additionally useful with untrimmed video data; indicating that a single
video incorporates a couple of subsequent things to do or several frames that are beside the
point to the ground-truth activities. Video cognizance algorithms face several challenges from
such data. Additionally, due to the fact feeding all of the frames to the studying mannequin
makes the computations very heavy, body choice is crucial for untrimmed video evaluation
displayed the area analysing as distinct MDPs, the place each and every one of them is
associated with a professional accountable for choosing one edge. Multi-agent reinforcement
studying solves the issue. The climate, for instance the video, is encoded with the aid of a
Convolutional Brain Organization (CNN) observed via an intermittent intelligence business
enterprise to seize the unique scenario data.
When there is no alternate in the movements of any of the agents, the episode is over. A
skilled baseline classifier's output is used to generate the reward. All the greater explicitly,
the prize is decided in mild of the distinction in the classifier's sure bet towards looking
forward to the proper class. The optimization manner makes use of the REINFORCE
algorithm.
Finding spatial consideration in the HAR task refers to the idea of keeping track of the edges'
most educational locations, which results in delivering more real recognition. Various
physical aspects of the body are required for many human tasks. Because of this, not all tasks
are now equally important to the skeleton data. This results in the introduction of a spatially
demanding interest discovery deep reinforcement learning-based strategy for selecting the
most informative joints in skeletal video frames. The joints used for reproduction might also
be anything from the body to the frame.
The method concerned with identifying the largest joints is planned as an MDP and is
resolved through support calculation. In this technique, the specialist,In Action 1, the
corresponding joint is again to the selected joint set if it used to be eliminated in the
preceding RL step, and if it is already protected in the chosen joints, it is removed. Action
zero suggests that no alternate need to be made. A pre-trained classifier that is related to the
reward characteristic utilized.
The majority of undertaking awareness strategies are supposed for third-person static digital
camera videos. There are no longer many offers with the recordings of first-person
perspective captured via wearable cameras. These kinds of recordings are egocentric records.
Processing high-resolution video data frequently requires significant computational
resources, according to DRL-based strategy is introduced to tackle this trouble in the subject
of exercise attention by way of selfish video which objectives monitoring down the locale of
hobby (return for cash invested) in every casing of the video.
A bounding container of a constant dimension is taken into consideration for the ROI, and the
RL hassle of discovering the nice place is modelled. An agent receives the contemporary
video frame as its enter state, and the motion includes two actual values defining the
horizontal and vertical moving quantity for the bounding field location. To maximize the
classifier's cumulative reward, the Actor-Critic technique is utilized. A deep community is
used to procedure the first-rate place of hobby after it has been identified, whilst a shallow
community is used to method the complete frame.
The consequences validated the way that they ought to genuinely decrease the computational
intricacy whilst protecting the exactness is any other approach for enhancing endeavour
cognizance costs that makes use of DRL to generate visible interest maps in video frames.
Through RL, they choose to create an interest procedure that is very comparable to how
human beings leap and see. Specifically, a CNN affords a REINFORCE interest agent with
an enter of a characteristic cube at every video frame. In order to come across a weight
distribution that identifies the area of hobby in every slice, this agent makes a predetermined
variety of jumps over the dice slices. An LSTM community will use the sum of the interest
maps observed in the slices for the remaining endeavour recognition. A comparable LSTM is
used to create the award which is framed by using thinking about each the volume of
authentic up-sides and bogus up-sides all collectively to examine excessive accuracy.
The consciousness project can't be carried out at the quilt of an pastime in some applications,
like video surveillance. As a result, there are some strategies that use movies of unfinished
duties to perceive an activity. Early undertaking awareness is the identify of this. A amazing
deal of workout routines are like every different at their early phases, so early motion
acknowledgment is notably hard and the great majority of the universal motion
acknowledgment techniques operate pitifully in such cases. In The creators introduced a
consistency rating which characterizes how correct the information succession of an
incomplete motion can be perceived.
Consequently, this strategy can also additionally be referred to as temporal interest finding.
After that, the chances received are sampled from a Bernoulli distribution to produce
movements with values of zero or 1. The undertaking indicates whether or not the side is
really helpful for the expectation capability of its previous casing either in succession (one),
or not (zero). The reward is designed in a way that results in the selection of the fewest
possible predictable frames. As a result, there are two components to the reward: The
accuracy of the classifier is one factor, and the cardinality of the chosen body set is another.
The agent is educated with the REINFORCE algorithm. That's what the creators reasoned the
consistency rankings are incredibly recommended to the early acknowledgment task.
Inspired by using human awareness behaviour, the authors of some other work in the DRL-
based early pastime attention subject hypothesized that except beside the point lessons all
through awareness can enhance performance. This is due to the fact of the way that the big
difference between the tremendous category and the terrible instructions is not the identical
for each poor class. As a result, the terrible training ought now not to be handled in the equal
way. A covering operation is used on the chance outputs of the classifier after it has been pre-
trained on all training in this paper.
Figure 2. Examples of using DRL for a) finding temporal attention, b) finding spatial
attention, c) early activity prediction, d) finding the best network architecture e) finding
fusion weights, f) optimizing the cluster centroids, and g) robot control, to improve the
human activity recognition performance
The cover determines which categories of information should be maintained and which ones
should be removed. Since there is no established ground truth for the optimum collection of
prohibiting classes, this method of thinking about the appropriate combination of negative
classes that should be excluded from the preparation set is seen as a support learning
problem.
The promise of the usage of neural networks as characteristic approximators for each the
country fee The action-value characteristic and feature V (s) Q(s, a) in visual-based RL tasks
again in the middle of the 2000s, when deep mastery research used to be progressing. The
studies that have combined deep neural networks with a reinforcement learning framework in
order to improve performance of studying manage insurance policies are mentioned in the
following sections, with an emphasis on these that are fed with uncooked enter records.
Neural Fitted Q mastering (NFQ), a model-free approach, has been proposed in [35]. NFQ
replace the hundreds of a multi-facet perceptron via RPROP calculation OF clump getting to
know method for making ready the talent networks which is particularly speedy in
examination with different directed inclining techniques, for relapsing the really worth
functionality the place fresh is executed disconnected. The update is based totally on a whole
set of transition experiences that have a triple structure (s, a, s), the places is the present day
state, a is the motion that was once selected, and s is the subsequent country that happens as a
end result of taking motion a. The transition experiences have already been gathered because
updating is done offline. In actuality, they are obtained through interaction with a real or
virtual environment. The suggested approach consists of two steps: (1) acquiring mastery set;
and (2) performing a clump update for preparing a multi-facet perceptron using RPROP
computation.
An strategy recognised as Deep Q getting to know Network (DQN) has currently been
developed via researchers working with DeepMind applied sciences [30]. DQN makes use of
the blessings of deep gaining knowledge of for summary illustration in studying most
excellent policy, which entails deciding on moves in such a way as to maximize the
anticipated fee of the complete rewards. It is an growth of the previous work Brain Fitted Q-
Learning (NFQ). The easiest reinforcement mastering approach (Q-learning) and a deep
convolutional neural community are mixed in DQN, permitting customers to play a range of
Atari 2600 video games completely by way of staring at the screen.
Up to this factor the proposed approach performed the first-rate non-stop specialists. On
many Atari video games with the identical community structure or hyperparameters, its
approach carried out higher than the human participant in some games. The widespread
outcomes have been the end result of a wide variety of elements that the preceding works did
no longer take into account [9]. First, traits in computing power, particularly the extremely
parallelized GPU technology, which has made it possible to train deep neural networks with a
variety of weight settings. Second, DQN has increased illustration learning by utilising a
massive deep CNN. Third, DQN has included trip replay for the related states issue.
However, in order for deep neural networks to examine higher representations and operate
well, adequate statistics have to be fed into the network. Because performing a giant range of
episodes to acquire samples is resource-intensive and even impossible, imposing this method
in a real-world surroundings like robotics is extraordinarily difficult and difficult.
A convolutional neural network's education facts are supplied by way of ability of the offline
Mont Carlo tree search planning. To be sure, they have fostered a few strategies which gain
from profound getting to know nets for conceptual portrayal and sans mannequin RL by way
of the use of UCT primarily based arranging method to produce enter data for the CNN.
The ALE framework is used as a testbed for the proposed techniques in this work as nicely as
in DQN. It performs higher than DQN in a wide variety of Atari 2600 games. UCT wishes a
lot of time between movements to get these results. Additionally, planning-based techniques
are inefficient for enjoying in actual time.
In order to enhance performance, the lookup in objectives to educate the grasp and manage
structures concurrently instead than separately. To attain this reason, getting to know a
approach which do each the perception and the manipulate mutually, they used profound
convolutional talent organizations. A PR2 robot's digicam offers the CNNs with uncooked
images, and the coverage is produced. Given the remark of the environment, the coverage is a
conditional Gaussian distribution that determines a likelihood distribution for actions. The
creators assessed their method with contrasting with different approach lookup processes on
specific undertakings, for example, striking a coat holder on a clothes rack, putting caps onto
pill bottles, building Lego bricks on a perfect platform, and so forth. also has demonstrated
essential results.
The improvement of an synthetic agent that can take part in a range of video games stays a
great AI challenge. Board video games like One of these challenging games is the two-player
game of Go, where the object is to encircle more territory than the opponent. Recurrent
neural networks in several dimensions (MDRNNs) and lengthy temporary reminiscence
(LSTMs) have been blended to create a approach for enjoying Go on small boards that the
authors describe in the function of MDRNN that approves it to use furnished data about the
two area dimensions of the sport board helps the proposed method.
In addition, The LSTM has been included into MDRNN to tackle the vanishing gradient
problem for RNNs. Model-free reinforcement research for POMDPs problems using policy
gradients with parameter-based exploration has beat evolution Strategies had been used to
teach the networks. Strikingly, as properly as have worried CNNs for enjoying Go sport
which enter facts used to be crude visible pixels. Their proposed techniques have come about
slicing part execution to the problem of waiting for the strikes made through grasp Go
players. To deal with the Go game, however, combining CNNs and RL frameworks may
additionally end result in extended performance.
The lookup makes use of end-to-end reinforcement gaining knowledge of to study most
efficient policies, simply like preceding works in the visible based totally RL domain. It used
a compacted repetitive talent networks which entails developmental calculations for
advancing the talent community as the endeavour esteem functionality inexact. It effectively
carried out two tough tasks, the visible Octopus Arm assignment and the TORCS race vehicle
using with high-dimensional visible facts streams.
Another region the place combining RL and deep getting to know methods can be most high-
quality is video prediction is one terrific piece. They have confirmed that their architectures
can extract spatial and temporal elements from Atari video games and generate 100-step
action-conditional future frames barring diverging.
RL and the unsupervised getting to know strategies of deep neural networks have been used
in a variety of tries to study representation. In the accompanying area, we will tackle a few
solo profound intelligence networks which are utilized to study minimized low-layered
encompass house of the RL task. Visual-based reinforcement gaining knowledge of duties
usually require two steps to complete. The first is, planning high-layered enter data into a
low-layered portrayal (which here, our core is making use of the solo mastering methods for
profound models). The 2d choice is to approximate the Q-value characteristic or manipulate
insurance policies with the aid of using an approximation approach on the discovered
compacted function space.
Profound Fitted Q-emphasis calculation has moreover been correctly used to end up
acquainted with the manipulate method for two manage errands, a shaft adjusting and a
hustling house vehicle, individually. They have observed two stages, (1) crude visible
statistics which had been caught by means of a computerized camera, is taken care of into an
auto-encoder community to reducing also, gathering the data country space, and (2) to
appraise the really worth capability, in preceding a bit based totally functionality
approximator has been utilized and in last, the Cluster RL method has been used. Be that as
it may, any functionality wager techniques can be utilized for attaining this step.
In most proper functions the Markov supposition that is not plausible, on the grounds that
proper states are simply rather recognizable and involving simply the current statuses for
impartial path may also now not precipitated arrive at the best procedure. POMDPs, in
distinction to Markov choice methods (MDPs), anticipate that the RL agent's enter states are
incomplete and can't incorporate all of the data required to pick the satisfactory subsequent
action. One way which can work with this irregularity is keeping the historic backdrop of the
previous perceptions, for this cause [30] have stacked a set of experiences of the ultimate four
strategies that the professional has as of late discovered in their examinations when they have
concerned Atari 2600 video games as the testbed.
When a device requires an arbitrary heritage of previous events, recurrent networks are
regularly used. RNNs are utilized as feature approximators, enabling them to produce
condensed characteristic areas of earlier located events. Indeed, the POMDP RL and DL
RNN mixture allows the agent to recall large preceding observations and lookup was once
one of the first to use RL in recurrent networks.
In the subject of robotics, a Long Short Term Memory (LSTM) recurrent neural network—a
specific sort of RNN that can examine lengthy time period dependencies of states that have
already been seen—provided the RL robotic with reminiscence capability. The Deep Q-
network has been modified with the aid of Hausknecht and Stone. so that it can be used in
environments the place observations may additionally be noisy and incomplete (such as
POMDPs environments). For this cause they have included a LSTM with a DQN.
DRL has been used for more than just finding attention, as shown in Table 1, including
determining the best network structure and the ideal weights for the merging of data
modalities. This abundance of endeavours might be seen as a pursuit problem in a vast area
where there is no ground truth. However, a reward signal can guide the lookup, making DRL
a useful choice. Due to its advantages over other techniques, such as their capacity to
represent continuous action spaces or their ability to converge more quickly, policy gradient
approaches have been predominantly used in DRL-based HAR systems. Even though these
approaches usually combine into a close-to-ideal configuration, they can nevertheless
accomplish the target objective and advancethe acknowledgment execution. The review
methods all have one thing in common.Feedback from a baseline recognition model that has
already been trained is used to determine the reward.
A new field of research called "human activity recognition" has evolvedin 2017 using deep
reinforcement learning. There are still loads of difficulties and open issues tobe tended to
from here on out. The following is a discussion of a few of these difficulties.
References
[Al-Amin et al., 2019] Md Al-Amin, Wenjin Tao, David Doell, Ravon Lingard, Zhaozheng
Yin, Ming C Leu, and Ruwen Qin. Action recognition in manufacturing assembly using
multimodal sensor fusion. Procedia Manufacturing, 39:158–167, 2019.
[Arulkumaran et al., 2017] Kai Arulkumaran, Marc Peter Deisenroth, Miles Brundage, and
Anil Anthony Bharath. Deep reinforcement learning: A brief survey. IEEE Signal Processing
Magazine, 34(6):26–38, 2017. [Dong et al., 2019]
Wenkai Dong, Zhaoxiang Zhang, and Tieniu Tan. Attention-aware sampling via deep
reinforcement learning for action recognition. In Proceedings of the AAAI Conference on
Artificial Intelligence, volume 33, pages 8247–8254, 2019.
[Gao et al., 2019] Yang Gao, Hong Yang, Peng Zhang, Chuan Zhou, and Yue Hu. Graphnas:
Graph neural architecture search with reinforcement learning. arXiv preprint
arXiv:1904.09981, 2019.
[Gowda et al., 2021] Shreyank N Gowda, Laura SevillaLara, Frank Keller, and Marcus
Rohrbach. Claster: clustering with reinforcement learning for zero-shot action recognition.
arXiv preprint arXiv:2101.07042, 2021.
[Guo et al., 2021] Jiale Guo, Qiang Liu, and Enqing Chen. A deep reinforcement learning
method for multimodal data fusion in action recognition. IEEE Signal Processing Letters,
2021.
[Haque et al., 2016] Albert Haque, Alexandre Alahi, and Li Fei-Fei. Recurrent attention
models for depth-based person identification. In Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, pages 1229–1238, 2016.
[Herath et al., 2017] Samitha Herath, Mehrtash Harandi, and Fatih Porikli. Going deeper into
action recognition: A survey. Image and vision computing, 60:4–21, 2017. [Imran and
Raman, 2020] Javed Imran and Balasubramanian Raman. Evaluating fusion of rgb-d and
inertial sen- sors for multimodal human action recognition. Journal of Ambient Intelligence
and Humanized Computing, 11(1):189–208, 2020.
[Jaafra et al., 2019] Yesmina Jaafra, Jean Luc Laurent, Aline Deruyver, and Mohamed Saber
Naceur. Reinforcement learning for neural architecture search: A review. Image and Vision
Computing, 89:57–66, 2019.
[Jaderberg et al., 2019] Max Jaderberg, Wojciech M Czarnecki, Iain Dunning, Luke Marris,
Guy Lever, Antonio Garcia Castaneda, Charles Beattie, Neil C Rabinowitz, Ari S Morcos,
Avraham Ruderman, et al. Human-level performance in 3d multiplayer games with
population-based reinforcement learning. Science, 364(6443):859–865, 2019.
[Ji et al., 2018] Yanli Ji, Yang, Xing Xu, and Heng Tao Shen. One-shot learning based
pattern transition map for action early recognition. Signal Processing, 143:364– 370, 2018.
[Klaser et al., 2008] Alexander Klaser, Marcin Marszałek, and Cordelia Schmid. A spatio-
temporal descriptor based on 3d-gradients. In BMVC 2008-19th British Machine Vision
Conference, pages 275–1. British Machine Vision Association, 2008.
[Kong and Fu, 2018] Yu Kong and Yun Fu. Human action recognition and prediction: A
survey. arXiv preprint arXiv:1806.11230, 2018.
[Kumrai et al., 2020] Teerawat Kumrai, Joseph Korpela, Takuya Maekawa, Yen Yu, and
Ryota Kanai. Human activity recognition with deep reinforcement learning using the camera
of a mobile robot.
[Laptev et al., 2008] Ivan Laptev, Marcin Marszalek, Cordelia Schmid, and Benjamin
Rozenfeld. Learning realistic human actions from movies. In 2008 IEEE Conference on
Computer Vision and Pattern Recognition, pages 1–8. IEEE, 2008.