Abysmal means bottomless; resembling an abyss in depth; unfathomable.
Perception is a procedure of acquiring, interpreting, selecting, and organizing sensory
information. Perception presumes sensing. In people, perception is aided by sensory organs.
In the area of AI, perception mechanism puts the data acquired by the sensors together in a
meaningful manner. Machine perception is the capability of a computer system to interpret
data in a manner that is similar to the way humans use their senses to relate to the world
around them. Inspired by the brain, deep neural networks (DNN) are thought to learn abstract
representations through their hierarchical architecture.
Deep learning is part of a broader family of machine learning methods based on learning data
representations, as opposed to task-specific algorithms.
Deep learning emerged in the last decade and extremely changed and is still changing our
current and future world. It refers to a ‘deep mining of data’ with so-called deep neural
networks: neural networks having so many layers. What a net is doing is cascading simple
linear transformations to represent highly non-linear functions that could efficiently
extract the basic structures and patterns within the data and map to an output of ‘making
sense of input’. Yes, neural Nets is a cascade of layers: Those are hidden: who knows
what is exactly happening! As one adds more and more ‘hidden’ layers, so the network gets
deeper and one makes it able to represent any function: they are universal approximators.
But getting deeper comes with a price: more layers mean more parameters to tune. Learning
millions of parameters requires big data, otherwise, neural networks will fail. The
learning/tuning process is a game of step back and forth in the space of numbers with a well
known back-propagation technique. This game is played in training the networks — just like
training a human which learns from his experience — well, mostly from his mistakes.
A type of neural layer is Convolutional Neural layer which turns a network to a
Convolutional Neural Network -- a neural network with particular ability to extract rich
contextual information from image-like data, mimicking how a human observer understands the
’seen’ world, by expressing it in terms of non-seen, non-attended, non-humanly-expressible
basic structures. How and why it performs far better than any other machine learning
techniques and continues to even beat human-level performance is a hot topic and several
technical proofs from optimization, probability, and statistics, mathematics, control
theory, etc. perspectives are available, but it is still a pipeline of linear
transformations, nothing more...
The work mostly shows the ‘hidden’ transformations happening in a network: summing and
multiplying things, adding some non-linearities, creating common basic structures, patterns
inside data. It creates highly non-linear functions that map ‘un-knowledge’ to ‘knowledge’.
As our time creates quintillions of bytes of information per day, how can we make sense of
this huge amount of data? Let networks do it four ourselves. We give all the human knowledge
and experience to the net then it will make sense of everything. As our life is becoming
non-sense, maybe we expect NN to learn and give us the sense of it.