The DSP is holding back the NeuroNet

So, on the most complicated #Machine Learning engines of #Humangnify, the DSP that is converting audio to Fourrier transform histogram is holding back the performance of the voice recognition. By restructuring the code a little, I think I can improve the speed and free it from the DSP gate. I really want to post all of the CoreML engines at one, I probably need one more day to get this done right.

Thanks for your patience.

Humangnify #DNLPs

#DLNPs are Deep Natural Language Processing Neuronal networks, we have written few of them, including some able to tell you the mood of people, this will allow you to get the temperature on people, based on your emails, recent audio interaction or social media account of this person. 

Out of those #DNLP, we are going to extract statistics about the people you interact, like how many time a person changed his mind, or contradicted you. We have good accuracy, but obviously, we have to keep working to make progress. If you join out sponsors, we will keep you at the edge of what is possible. Some of those discovery are in Patent Pending phase.

iPhone XsMax, got it!

So, one of the best workload we have in Humangnify is the DeepSpeech port, it was rewritten to allow the Apple DSP to do the signal processing, instead of doing the fourrier transform by hand … Unfortunatly, the speed on the new iPhone Xs Max does not seem to improve because the DSP is holding back the rest, so, I need to rebalance a little the algorithm, to allow more room for the Neuronal part. There is no point to benchmark something with it not being the bottle neck, so, I will be coding it differently tomorrow. 

The Benchmark pieces

???? Using Text data input.

???? Using audio as input.

????️ Using picture or video as input.

Restnet50????️

ResNet is a short name for Residual Network. As the name of the network indicates, the new terminology that this network introduces is residual learning.

What is the need for Residual Learning?

Deep convolutional neural networks have led to a series of breakthroughs for image classification. Many other visual recognition tasks have also greatly benefited from very deep models. So, over the years there is a trend to go more deeper, to solve more complex tasks and to also increase /improve the classification/recognition accuracy. But, as we go deeper; the training of neural network becomes difficult and also the accuracy starts saturating and then degrades also. Residual Learning tries to solve both these problems.

What is Residual Learning?

In general, in a deep convolutional neural network, several layers are stacked and are trained to the task at hand. The network learns several low/mid/high level features at the end of its layers. In residual learning, instead of trying to learn some features, we try to learn some residual. Residual can be simply understood as subtraction of feature learned from input of that layer. ResNet does this using shortcut connections (directly connecting input of nth layer to some (n+x)th layer. It has proved that training this form of networks is easier than training simple deep convolutional neural networks and also the problem of degrading accuracy is resolved.

source

Gender Net ????️

Extract age and gender of a face.

Car Recognition ????️

recognize cars model and brand in USA market.

CNNEmotions ????️

Part of GoogleNet, this one recognize emotion out of a picture and provide an idea of the emotion in the picture.

Tweet Classifier????

This one classify how people feel, and what kind of emotions they try to express in a TEXT, using #DNLP.

GoogleNet Places????️

Recognize places out of a picture, it is part of GoogleNet

Nudity Decency????️

This one give you a score of how appropriate this picture is regarding to nudify.

Face Emotions ????️

This one does recognize face emotions, using a different structure of network,  I use it to validate the results of CNNEmotions.

Inception V3 ????️

A legend in picture machine learning, it does qualify objects in the picture, it is never too far from the best of art. This version comes from googleNet.

Sentiment Polarity ????

It is a #DNLP, tells you , out of your texts, the sentiment convey by a sentence.    


Tweet Personal Attack detector????

Detect agreesivity toward the reader, in House developped.

DeepSpeech Voice Recognition ????

The famous DeepSpeech, nodified and recoded to work with iPhone DSP and Apple convolution neuronal network engine. It recognizes pretty accurately what you say, while not buring your battery down fast. Tuned in house, and used to index what ever you want, your meeting, your day, your life …

Speaker Recognition ????

Detect when the people speaking change to an other person, learn the new person on the fly, and then, generate a time table of different speakers in a recorded audio. It allows humangnify to provide you a perfectly indexed meeting.

Insult Detector ????

This one is trivial, it detecting insults, it is home made, using a regular #DLNP, it takes a text and score it from 0 to 100, zero is polite text, 100 is seriously deep disgusting bad English language.

Before you start looking at the numbers…

So, if you are here, you are a geek, you seek better understand of a tech topic. This is going to be my opinion, based on the data I collected, with the best of my ability, I have 20 years of working in Benchmarking, but, I am only 1 year old on iOS, so, if you see any non accuracy on the iOS topic, please feel free to contact me on twitter and debate what I have done wrong. My tweeter is @fpiednoel. Now, that we all understand let’s get to the meat of this page, the performance of the Apple devices on iOS and MachineLearning, and what does it means for consumers, and the future of Machine Learning. I forgot, this place is supposed to be an efficient way to communicate, not a professional site, at Humangnify, we focus on the MachineLearning, not the pretty stuff.