audioVersionDurationSec
float64
0
3.27k
codeBlock
stringlengths
3
77.5k
codeBlockCount
float64
0
389
collectionId
stringlengths
9
12
createdDate
stringclasses
741 values
createdDatetime
stringlengths
19
19
firstPublishedDate
stringclasses
610 values
firstPublishedDatetime
stringlengths
19
19
imageCount
float64
0
263
isSubscriptionLocked
bool
2 classes
language
stringclasses
52 values
latestPublishedDate
stringclasses
577 values
latestPublishedDatetime
stringlengths
19
19
linksCount
float64
0
1.18k
postId
stringlengths
8
12
readingTime
float64
0
99.6
recommends
float64
0
42.3k
responsesCreatedCount
float64
0
3.08k
socialRecommendsCount
float64
0
3
subTitle
stringlengths
1
141
tagsCount
float64
1
6
text
stringlengths
1
145k
title
stringlengths
1
200
totalClapCount
float64
0
292k
uniqueSlug
stringlengths
12
119
updatedDate
stringclasses
431 values
updatedDatetime
stringlengths
19
19
url
stringlengths
32
829
vote
bool
2 classes
wordCount
float64
0
25k
publicationdescription
stringlengths
1
280
publicationdomain
stringlengths
6
35
publicationfacebookPageName
stringlengths
2
46
publicationfollowerCount
float64
publicationname
stringlengths
4
139
publicationpublicEmail
stringlengths
8
47
publicationslug
stringlengths
3
50
publicationtags
stringlengths
2
116
publicationtwitterUsername
stringlengths
1
15
tag_name
stringlengths
1
25
slug
stringlengths
1
25
name
stringlengths
1
25
postCount
float64
0
332k
author
stringlengths
1
50
bio
stringlengths
1
185
userId
stringlengths
8
12
userName
stringlengths
2
30
usersFollowedByCount
float64
0
334k
usersFollowedCount
float64
0
85.9k
scrappedDate
float64
20.2M
20.2M
claps
stringclasses
163 values
reading_time
float64
2
31
link
stringclasses
230 values
authors
stringlengths
2
392
timestamp
stringlengths
19
32
tags
stringlengths
6
263
0
null
0
null
2017-12-14
2017-12-14 07:38:27
2017-12-17
2017-12-17 22:55:59
4
false
fr
2017-12-17
2017-12-17 22:59:49
3
1339056062fa
2.9
1
0
0
C’est une constante dans la mise en place d’une optimisation de processus par IA: Quand manuellement et sur un petit échantillon, les…
5
AI : Comment monter en puissance ? Robotiser le tagging des objets dans les images fixes C’est une constante dans la mise en place d’une optimisation de processus par IA: Quand manuellement et sur un petit échantillon, les résultats sont là, il faut réfléchir de suite à la manière de “scaler” rapidement. Soit vous avez un fond déjà qualifié et vous considérez qu’il est exhaustif, alors la mise en place sera rapide car basé sur ce fond (et elle ne bougera pas). Soit vous n’avez pas de fond qualifié ou bien vous savez que l’organisation devra être apprenante, dans ce cas — et en fonction des volumes à venir — vous devez avoir la capacité d’être productif dans l’alimentation de l’apprentissage supervisé. Nous avons pu le voir dans l’article précédent (https://medium.com/@aescart1/la-computer-vision-d%C3%A9tection-dobjet-avec-tensorflow-5ee819d74f61), ce qui est fastidieux, c’est le fait de devoir travailler chaque image une par une. Même avec des outils optimisés, c’est très chronophage (Il faut compter 30 min de tag manuel pour 100 images environ) Par contre, nous avons vu dans le premier article — exemple de la montre (https://medium.com/@aescart1/la-computer-vision-touche-tous-les-domaines-de-lentreprise-partie-1-13d14464a30b) — que l’on pouvait partir d’une image pour la retrouver dans une autre. C’est ce principe que l’on va appliquer ici. Une fois qu’on a lancé le process de création d’image fixe à partir d’une vidéo (Cf. article précédent — video_2_image.py), il faut ainsi : Sélectionner une image nette et de face, afin qu’elle serve de source au tagging Modifier cette image afin qu’elle ne représente plus que la face à rechercher Lancer la recherche de cette source dans toutes les images, en extraire des coordonnées des extrémités et générer le fichier xml de tag. Vérifier le résultat pour pouvoir et lancer l’apprentissage supervisé Optionnel : avant la génération du XML, on peut regarder si l’aire trouvée se trouve dans la moyenne et si ce n’est pas le cas, mettre l’image de côté. Prenons le cas de la vidéo suivante. Quand on la décompose, on se retrouve avec les fichiers suivants (cela nous ferait 265 images à traiter) Il faut donc trouver la bonne image, de face, et n’extraire que cette face avec un fond transparent. Cette face permettra la recherche dans toutes les images pour faire le tagging. Le fond transparent (format png) permettra de ne pas perturber le tagging. On utilisera donc OpenImaj pour rechercher la source dans les cibles. Le code est très simple, on recherche tous les points de concordance et on ne garde que les plus extrêmes afin de former un carré. Le code Java simplifié est à l’adresse suivante : https://gist.github.com/aescart1/aebf625333422eba456b7fd244814c3c Quand on le fait tourner, on obtient les fichiers XML pour chaque image. Si on inspecte le tagging en lui même, on s’aperçoit que le résultat est très acceptable. Une image taggué automatiquement par le programme. Le résultat est similaire à un tagguing manuel. Voici une vidéo qui présente quelques apprentissages. On pourra imaginer monter en puissance en traitant plusieurs dossier, en générant le .pbtxt et en lançant automatiquement l’apprentissage. Ce qui est important, c’est que le travail manuel s’en retrouve très simplifié et qu’on a donc la possibilité de faire du deep learning supervisé, certes, mais moins :) Pour conclure: On connaît les cas métiers où la Computer vision peut automatiser les process On sait alimenter et mettre en production un réseau de neurones pour automatiser ces process On sait automatiser l’alimentation du CNN afin de réduire le time to market Cela autorise de l’ambition dans la mise en place des projets d’IA autour de la “computer vision” et de la détection d’objet.
AI : Comment monter en puissance ? Robotiser le tagging des objets dans les images fixes
1
ai-comment-monter-en-puissance-robotiser-le-tagging-des-objets-dans-les-images-fixes-1339056062fa
2018-04-10
2018-04-10 20:34:59
https://medium.com/s/story/ai-comment-monter-en-puissance-robotiser-le-tagging-des-objets-dans-les-images-fixes-1339056062fa
false
583
null
null
null
null
null
null
null
null
null
Image Recognition
image-recognition
Image Recognition
517
Aurélien Escartin
Entrepreneur in the retail revolution | liveshop.ai | Living the AI & Digital Joy #AI #UX/UI #Tensorflow #python
be2440d383cf
aescart1
56
109
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-11-08
2017-11-08 01:55:07
2017-11-12
2017-11-12 06:33:20
1
false
en
2017-11-12
2017-11-12 06:35:39
4
13399dbd0f3e
2.064151
0
0
0
I just finished a book and the last words were:
4
Day 2/28: My Journey to understanding Artificial Intelligence Is AI coming for your job? Dialogue between Yann LeCun and Matissa Hollister, Oct 11th 2017, Montréal I just finished a book and the last words were: “There are three types of people: those who make things happen, those who watch, those who have no idea what happened. The choice is yours. Never too late to catch the train”. They’ve stuck with me ever since. It’s my second day writing this journal for a class at McGill University. We are asked to pick a topic and explore its implications in retrospect to societal themes discussed in class. I’ve picked the topic of Artificial Intelligence, and every day, for 28 days, I will be posting my thoughts from books, articles, conversations with experts, comments, etc.. And so it starts…I’ve just caught the train and I am on a journey to be apart of ‘those who watch’. My starting point is: Artificial Intelligence is bound to disrupt the workplace. To ensure positive societal development a system of ethics must be developed. AI, understood by the few, will increase inequalities. What is Artificial Intelligence ? I understand AI as being a process by which a machine is capable to imitate intelligent human behaviours. Why all the hype around Artificial Intelligence? I study in Montréal and the city is becoming a hub for AI. I’ve subsequently developed an interest in the field. Attending sold out AI events, like the Google Brain talk or Yann LeCun’s intervention, has made it apparent that I am not the only one curious of its developments. And rightly so, considering the speed of its development! So, What’s the big deal? Its applications concern all of us; AI is applicable in every industry from science, to sports, to music, to finance, and the list goes on. AI is said to be the biggest change since the ‘Industrial Revolution’, where step-by-step it will transform industries for the better or for the worse. It is for us to decide. Will AI take over humans ? Meh… Singularity (the point at which AI will surpass human intelligence) is far far away. But not something to ignore, Masayoshi Son, founder & CEO of Softbank, recently announced a $100 billion fund to support his 30 year vision plan. He believes in 30 years time, singularity will be achieved. Will AI takeover our jobs ? Yes, many. AI is increasingly taking translator jobs, as its applications are getting better and better. If you compare google translate today to what it was 3 years ago; whilst there still is some work to do, it is much better today. AI will be taking over the job of drivers with autonomous cars — when? remains a question of essence — especially after the first crash in Vegas on the first day of launch. Thank god, no one was injured! Its late gotta catch some sleep...zzzz See you tomorrow morning! My last review of AI was quite pessimistic, I’ll reverse things around tomorrow, and try speak of it positively. Cheerio :)
Day 2/28: My Journey to understanding Artificial Intelligence
0
day-2-28-my-journey-to-understanding-artificial-intelligence-13399dbd0f3e
2018-02-15
2018-02-15 10:54:29
https://medium.com/s/story/day-2-28-my-journey-to-understanding-artificial-intelligence-13399dbd0f3e
false
494
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Augustin Mallon
Student @McGill Business | Entrepreneur | Traveller | China, US, Israeli, Europe tech valleys | AI, Blockchain, IOT | 🌏💫
d4271ea5d190
AugustinMallon
6
7
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-13
2018-07-13 21:14:19
2018-07-13
2018-07-13 21:26:15
2
false
en
2018-07-13
2018-07-13 21:26:15
0
133bb3d1f5e6
2.896541
1
0
0
Artificial Intelligence (AI), the ability of a machine or a computer program to think and learn like a human, for decades was more fiction…
5
The Promise of AI Artificial Intelligence (AI), the ability of a machine or a computer program to think and learn like a human, for decades was more fiction than reality. But today it’s impacting nearly every industry — from agriculture to manufacturing, healthcare, fintech and insurance, cyber security, retail, travel, sports, logistics, marketing and more. And the global race is on to take the lead in AI. In fact, China last summer unveiled its plan to become the world leader in AI and create an industry worth $150 billion to its economy by 2030. Technologists here see this as a direct challenge to America’s lead in arguably the most important tech research to come along in decades. We’re already using AI, interacting with technology in new ways — from giving voice commands to washer-dryers to playing advanced gesture-controlled video games. What does the future hold? Here are a few trends to mull over, some in the near future, some down the road: · Consumers will become even more comfortable with voice-based interfaces (Amazon’s Echo, Google Home, Apple’s Airpod), as smart assistants become integrated into computers, smartphones and even televisions. · Robots will be doing more of the heavy lifting with human workers taking over high-end jobs including robot maintenance and operation. This means warehouses and logistics will become more efficient. Amazon already uses over 100,000 robots in various warehouses, but at the same time the company is creating thousands of new jobs for humans in its new fulfillment centers. · AI is giving surveillance cameras digital brains to match their eyes, letting them analyze live video with no humans necessary. This could be good news for public safety, assisting police and first responders to better spot crimes and accidents; what’s more, these brains have a range of scientific and industrial applications. However, it also raises serious questions about the future of privacy. · Experts predict that, within the next decade, AI will outperform humans in relatively simple tasks such as translating languages, writing school essays, and driving trucks. More complicated tasks like writing a bestselling book or working as a surgeon, however, will take machines much more time to learn. AI is expected to master these two skills by 2049 and 2053 accordingly. Brands like USA Today, CBS and Hearst are already using AI technology to generate content. · The Defense Advanced Research Project Agency (DARPA) is working with a company to develop a series of robots designed for “disaster relief,” though the technology could be used in a combat role as well. On the marketing front, which is in our wheelhouse, marketers are looking at how AI can help social media advertisers, such as ways to enhance business intelligence, marketing research and forecasting accuracy; generate leads and acquire customers; enrich the customer experience with AI-enhanced call center technology, bots and virtual digital assistants, smarter search interfaces and recommender systems that help with different types of customer support; and content creation and dynamic creative to generate on-demand ads, articles, summaries, promotional materials, etc. The Bottom Line on AI A survey conducted in late 2017 of senior executives across all industry sectors on the role of AI revealed some interesting takeaways (Marketing Professionals): · 64% say AI will eventually allow staff to perform more varied roles by enabling employees to find solutions to problems that would previously have been referred up the chain of command · 79% say it will be standard practice to use AI to schedule/coordinate work within five years · 76% say AI will be used as part of quality control within the next five years · 70% of executives expect robotic (AI-driven) automation to principally replace or support human workers in administrative roles within 20 years, with some 41% expecting robotic automation to principally replace or support human workers in customer-facing roles within 20 years. How are you using AI to impact your industry? Sources: Marketing Professionals; Future of Humanity Institute, Oxford; CB Insights’ Verge
The Promise of AI
2
the-promise-of-ai-133bb3d1f5e6
2018-07-13
2018-07-13 21:26:15
https://medium.com/s/story/the-promise-of-ai-133bb3d1f5e6
false
666
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Stellar Agency
Stellar is a digital design agency focused on delivering world-class interactive experiences and products. Face The Future.
2d8f9a157a48
stellaragency
4
3
20,181,104
null
null
null
null
null
null
0
null
0
16de5c9bea5a
2018-04-10
2018-04-10 16:04:24
2018-05-02
2018-05-02 16:15:05
15
false
en
2018-06-13
2018-06-13 17:29:23
18
133bd6fb568e
12.303774
30
2
0
Written by Thibault Gisselbrecht and Joseph Dureau
5
ERROR: type should be string, got "https://www.pinterest.com/pin/402368547950165147/\nMachine Learning on Voice: a gentle introduction with Snips Personal Wake Word Detector\nWritten by Thibault Gisselbrecht and Joseph Dureau\nThe wake word is the word you use to start a conversation with a voice assistant. Inside every device in which a voice assistant is embedded, a tiny process keeps listening, waiting to detect the wake word out of a continuous stream of audio. The process at stake is both typical of how voice is generally transformed to feed a machine learning model, and quite simple from a Machine Learning perspective. We are going to take this example as an illustration of how Machine Learning is done on voice.\nThe personal wake word detector is a new feature of the Snips Voice Platform, that we release in response to strong demand from the community. It makes it possible for anyone to pick any wake word they want to use to call their voice assistant. Go straight to the tutorial if you want to start playing with it. To understand how this detector works, and how voice is handled in many other Machine Learning applications, carry on reading!\nPersonal and Universal Wake words\nThere are two kinds of wake word detectors: Universal and Personal ones.\nThe universal wake word detectors are trained on a large variety of voices. The underlying model is generally a Deep Learning algorithm, that is trained to identify when anyone says the wake word.\nOn the other hand, the personal wake word detector is trained locally, on your device, with a small number of voice samples that you provided. This alternative is much more versatile compared to the universal one, since it allows you to use any arbitrary wake word. The difference is just that it’s not meant to work when someone else says the wake word.\nWhat’s common to both type of detectors is the way voice is transformed before it’s fed into the Machine Learning model. This pre-processing is actually shared with many other applications of Machine Learning with voice, like Speech Recognition, or Speaker Identification. It just turns out that results obtained with this pre-processing step are often better than without, although exceptions are starting to arise.\nLet’s see how this personal wake word detector works in terms of training, inference, and performances.\nTraining\nAudio trimming\nThe principle of the personal wake word detector is to compare an incoming stream of audio to a set templates of the wake word recorded by a user. It is a nearest neighbor logic. Hence, the first step is to acquire those templates. In practice, since we have no idea of how long the wake word will last, a margin is taken to record them. In our case, we give the user 2 seconds to record each template (see the documentation here). The sample rate used to acquire the sound is set to 16000 samples per seconds. Each record is then initially composed of 32000 samples.\nNaturally, what comes before and after the wake word in the recording is not useful. If the recording takes place in a quiet environment, which we heavily recommend, what comes before and after are silences. In order to remove those silences and keep only the meaningful part of each template, a process called trimming is applied. This process consists in:\nDividing the signal into small chunks (framing)\nComputing the energy of each frame\nRemoving every frame which energy is lower than a predefined threshold.\nFor example, a 65 milliseconds signal might be sliced into 5 frames of 0.025 seconds with an overlap of 0.01 seconds between each frames. The figure below illustrates the framing process on a short audio signal.\n5 frames of 0.025 seconds with 0.01 second overlap\nThe energy of a frame is the mean of the squares of the signal. A classic approach is to compute the energy of each frame, and compare it to a predefined threshold to classify the frame as silence or not.\nTo increase robustness, we take a slightly different approach. We set a threshold on the ratio of energies between the energy of each frame, and the energy of the frame with highest energy. If this ratio is below 20 decibels, the frame is classified as silence. The threshold can be manually configured in the Snips platform.\nThe trimming process is illustrated below. On the left, the signal was recorded in clean conditions (no noise), and the trimmed signal strictly captures the wake word. On the right, an example with background noise recorded after the wake word. This example shows how noise can disrupt the trimming process.\nTrimming process in various recording conditions\nFeature extraction\nThe second step of the training process is to extract a concise and meaningful representation of each template, to feed a simple machine learning model. This process is called feature extraction.\nMel Frequency Cepstral Coefficients (MFCCs) are widely used for Automatic Speech Recognition applications. This transformation tries to mimic some parts of the human speech perception by reproducing the logarithmic perception of loudness by the human ear. For a full tutorial on how MFFCs work, follow this link. The process to obtain the MFCCs from an audio signal is the following:\na. Pre-emphasis (optional). This step is aimed at amplifying the high frequencies, in order to balance the signal and improve the overall signal to noise ratio. In practice, we use a pre-emphasis coefficient if 0.97.\n\nb. Framing. The signal is splitted into fixed-sized slices (usually between 20 ms to 40 ms long), with a predefined amount of overlap between each slice. This step is the same than the one described in the initial Trimming step, but the window sizes and the overlaps might be different. Let us denote nFrames the number of frames obtained from the original signal.\nc. Windowing. A Povey window function is applied to each frame to reduce side lobes. The Povey window function is similar to the Hamming window, but equals zero at the edges.\nFrame windowing.\nAt this point the original has been pre emphasised, framed and windowed, which we illustrate in the figure below.\nEmphasised, framed and windowed signal.\nd. Transformation. We first compute the Discrete Fourier Transform for each frame, with a predefined number of components (512 here). Then, we compute the square of the module of each coefficient to obtain the energy distribution of the signal across frequencies. At this stage, each frame is represented by 512 values, which represent the energy of the signal at different frequencies. The i-th value of this signal corresponds to a frequency of i*16000/(2*512).\nWe illustrate this on our toy examples below.\n\ne. MEL scale mapping. We map each component onto the mel scale using triangular filters: this transformation is aimed at reflecting the way the human ear perceive sounds. The number of filters, noted nFilters, is a parameter of the algorithm. At that point, each frame comes with a vector of size nFilters. We then take the log of each component to obtain the log-filter banks.\nf. Normalisation. Finally, we apply a discrete cosine transform (DCT) and normalize each component by removing their mean. This last step is meant to decorrelate the log-filter bank.\nAt this point, our audio sample has been transformed into a feature matrix of size nFrames ×nFilters where nFrames is the number of frames resulting from the second step above and nFilters is selected the number of mel coefficients at the last step. The figure below illustrates the feature matrix of the trimmed audio signal from the previous section.\nFeatures extraction process\nIn the next section, we will explain how we use those transformed templates in our wake word detection algorithm.\nInference\nDynamic Time Warping\nNow that we have a clean representation of the audio templates, our goal is to compare an incoming stream of audios to these templates, in order to decide whether a wake word is present or not is the audio stream.\nTo make things simpler, let us consider for now that only one audio template is available. Concretely, the live audio stream is sliced into windows with a size equal to the template and a defined overlapping shift. Each window is then processed with the feature extractor, and finally compared to the template. To perform this comparison and output a decision, we use Dynamic Time Warping, which measures the similarity between two time series (see this link for a full introduction to DTW techniques).\nLet us consider two times series of respective sizes N × K and M × K, where N and M are time dimensions, and K is the dimension of the features space. In our application N=nFrames[Template], M=nFrames[Stream] and K=nFilters.\nThe first step to compute the DTW between those series is to evaluate the related pairwise element cost matrix. The metric used to compute this matrix can change depending on the need, for our algorithm we chose the cosine similarity. Each element of this matrix is defined as:\n\n\nAfter this step we end with a matrix of size (N,M) called cost matrix. The aim of DTW is to find a path going from the first element (with coordinates (1,1)) to the last element (with coordinates (N,M)) of the cost matrix with a minimal cumulated cost. This path is called an optimal warping path and its total cost defines the DTW distance between both signals.\nThe figure below on the left represent the cost matrix between two series together with the related optimal warping path. Intuitively, the warping path tends to align the points of the two time series together with the constraint that every point from sequence 1 (respectively sequence 2) must be mapped to at least one point of sequence 2 (respectively sequence 1). In our specific case, since a wake word is always pronounced at a relatively constant tempo and since all its syllables are always pronounced in the same order, some warping path can be discarded. This constraint can be encoded directly encoded in the cost matrix by setting infinite cost to each element in the area that we want to discard. For our wake word detection algorithm we use a diagonal constraint (see figure below on the left).\nDynamic Time Warping\nReference distance and prediction\nAt this point, we are able to compare a live audio stream with each template by computing the corresponding DTWs. Formally, if we have 3 templates and consider a window from the audio stream, we can compute the 3 related DTWs, denoted DTW₁, DTW₂ and DTW₃. In practice, the duration of the audio stream window is set to the average duration of the 3 templates.\nOur objective is to classify the audio stream window as containing a wake word if its DTW with respect to at least one the templates is less than a predefined threshold. An optimal value for this threshold can be found empirically for each specific wake word. Yet, this value will heavily depend on the size of the templates. The higher the number of frames, the longer the path, and hence the highest the DTW. To counter this effect, and to be able to define a universal threshold (as in not wake word dependent), each DTW is normalized by the sum of the temporal dimension of both inputs.\nFinally, if we call the decision threshold 𝜏, and consider an input sequence from the audio stream, the detector will trigger if either of the normalised versions of DTW₁, DTW₂ and DTW₃ is lower than 𝜏.\nLet’s now see how we determined a good default value for 𝜏.\nConfidence\nThe wake word detection problem can be seen as a binary classification problem. The audio stream window either contains, or doesn’t contain the wake word. To tune this kind of problems, it is particularly useful to be able to output a probability to belong to each class (in our case to be or not to be a wake word). To this aim, we artificially define 3 confidences related to each input sample:\n\nwhere i is template indice. Note that if DTWᵢ is less than 𝜏, then the probability will be greater that 0.5, and will increase as DTWᵢ becomes smaller, as expected. Finally, the probability threshold above which the detector will trigger is the parameter that we expose to the user (see the tutorial here). We call it the sensitivity. The higher sensitivity, the higher the number of false alarm and the lower the number of missed wake words.\nThe next section is dedicated to an analysis of the performances of our algorithm, and its limitations.\nPerformances and limits\nFinding an acceptable reference distance 𝜏\nWe looked for an acceptable universal value for 𝜏. To do so, we empirically recorded different wake word templates, for different people, languages, and lengths of the wake word. For each of them, we computed the normalized DTWs with respect to the following set of audio samples:\nOther templates, not noisy: for each wake word, and each template of this wake word, this is the set of all the other wake word templates of the same wake word, except the one in question. Ideally, all DTWs with respect to the studied template should be lower than 𝜏.\nOther templates, noisy (20db): similarly, for each template of each wake word, this is the set of templates of the same wake word, except the one in question, augmented with background noise, with a signal to noise ratio of 20 decibels. Ideally, all DTWs with respect to the studied template should be lower than 𝜏.\nVoices: this is a set of audio recordings composed of people pronouncing random text. This set is around 10 hours long in total. Ideally, all DTWs with respect to the wake word templates should be greater than 𝜏.\nNoises: this is a set of recordings of background noise . Ideally, all DTWs with respect to the wake word templates should be greater than 𝜏.\nWe repeated this process for each wake word template, for different wake words, and aggregated the results. The figure below represents the normalized DTW distribution for each of the sets defined above. Looking at this figures gives a first intuition about the distance threshold that could be used. It appears that setting it around 0.22 will lead to a good separation between wake word and non-wake word recordings, both with and without background noise.\n\nTo confirm this intuition we computed the False Alarm rate per Hour (FAH) on the Voices dataset, that is the trickiest one, and the False Rejection rate (FRR) for actual templates, both in noisy and non noisy conditions, with a distance threshold of 0.22. The FAH quantifies the number of times the detector will accidentally trigger each hour while the seconds quantifies the rate at which a wake word will be missed. Both measures are the most commonly used in the literature for this task.\nOn voices, the False Alarm rate per Hour is 2.06, which means that with people constantly talking next to it, the wake word Detector will trigger about twice per hour with default settings. Improved performances can be obtained by tuning decision parameters on a case by case setting. The False Rejection rate is 0% without background noise, and 0.3% in noisy conditions.\nLimits of the approach\nTo push the exploration further, we increase the level of noise in the test datasets. We created a first dataset with 10 decibels signal to noise ratio, which is lower than the 20db used initially, and a second one at 5 decibels signal to noise ratio. Intuitively, the performance of the algorithm should decrease when the signal to noise ratio decreases, since the Wake word becomes more difficult to detect.\nThe same plots as in the previous section are shown with the 10 decibels set on the left and the 5 decibels on the right. As expected, in both cases the frontier between positive and negative samples becomes more blurry as the signal to noise ratio decreases.\n\nKeeping the distance reference to a value of 0.22 the False Rejection rate respectively increase to 2.8% and 20.4% with 10 and 5 decibels signal to noise ratios.\nOf course, as mentioned before, those are average performances obtained with default sensitivity threshold. Improved performances can be obtained by adapting the sensitivity for each Wake word.\nIn order to improve the robustness of the system, we are currently working on two approaches:\nAdding an audio frontend in order to artificially reduce the ambient noise (noise cancellation)\nDeep learning approaches that have been shown to be more robust to noise and transfer knowledge between all Wake words.\nTry it out!\nNow that you understand how sound is cleaned, transformed and processed to determine if a wake word has been said, try it out! A full tutorial is available here. To experiment further, you will be able to play with all the parameters defined in this article for trimming, feature extraction, etc in the script_recording.py file.\nHappy hacking, and don’t hesitate to share your feedback and new ideas!\n\nIf you liked this article and care about Privacy, smash that clap button, then tweet everyone 👉👉 tibo_gissel, jodureau, and snips!\nIt is also highly likely we have a job for you at Snips 🤩! We are the largest voice startup in Europe, and are hiring in machine learning, software engineering, blockchain, sales, product, marketing, etc…\n"
Machine Learning on Voice: a gentle introduction with Snips Personal Wake Word Detector
181
machine-learning-on-voice-a-gentle-introduction-with-snips-personal-wake-word-detector-133bd6fb568e
2018-06-13
2018-06-13 17:29:24
https://medium.com/s/story/machine-learning-on-voice-a-gentle-introduction-with-snips-personal-wake-word-detector-133bd6fb568e
false
2,863
This publication features the articles written by the Snips team, fellows, and friends. Snips started as an AI lab in 2013, and now builds Private-by-Design, decentralized, open source voice assistants.
null
snipscompany
null
Snips Blog
contact@snips.ai
snips-ai
ARTIFICIAL INTELLIGENCE,PRIVACY BY DESIGN,VOICE ASSISTANT,INTERNET OF THINGS,PRIVACY
snips
Machine Learning
machine-learning
Machine Learning
51,320
Thibault Gisselbrecht
Senior machine learning scientist @snips
2c57822545b7
thibault.gisselbrecht
29
8
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-09-27
2017-09-27 04:50:16
2017-09-27
2017-09-27 04:52:19
2
false
en
2018-08-30
2018-08-30 18:44:55
14
133d4382b6d5
1.92673
0
0
0
Yoshua’s new paper essentially deals with discussing non-trivial priors in some learning model, i.e. babies’ brains are pre-equipped with…
1
[Article.Ai] Freaky — why Yoshua Bengio’s new “Consciousness Prior” paper is freaky to me Yoshua’s new paper essentially deals with discussing non-trivial priors in some learning model, i.e. babies’ brains are pre-equipped with some “physics priors”, namely the ability for babies to intuitively know laws of physics, pertinently while learning by reinforcement. It is palpable that the phrasing “intuitively know laws of physics” above, should not be confused for nobel laureate or physics undergrad aligned babies that write or understand physics papers/exams; instead, the aforesaid phrasing simply conveys that babies’ brains are pre-baked with ways to naturally exercise physics based expectations w.r.t. interactions with objects in their world, as indicated by Aimee Stahl and Lisa Feigenson. I won’t go into mathematical detail regarding Bengio’s paper here, but you may better understand Bengio’s paper by viewing a paper of mine written last year. (A clear overview is found here). Although the aforesaid overview entails the complex topic of supermanifolds, and thus somewhat goes beyond manifolds as Bengio’s paper entails, the overview is quite clear, and may help readers here to better digest Bengio’s paper. ▼ ▼ On a separately fun note, I shall underline after the meme below, why Bengio’s paper is quite freaky to me. When I read the abstract, I quickly thought that Bengio was talking about learning some laws of physics, in conjunction with RL (reinforcement learning). Surely enough, reading the entire paper he does mention babies and intuitive physics, and RL. ▽ ▽ Last year I wrote a paper called “Thought Curvature”, about utilizing Supermathematics to learn laws of physics in tandem with RL (reinforcement learning) in the Supermanifold regime. (See a clear overview here) ▼ ▼ Some differences are: (a) Unlike Bengio’s paper, my paper presented a somewhat thorough way to do an experiment to test for the viability of the somewhat novel structure introduced in my paper. (b) My papers lack the entire description about “mapping and verifying conscious actions in culture” via language, whereby my papers don’t intend to describe a framework for consciousness. ▽ ▽ ▽ Footnote This is perhaps promising news though, it’s freaky to me as I didn’t think my papers were arXiv worthy yet, but Yoshua’s presentation here is perhaps changing my mind. Supermathematics-and-Artificial-General-Intelligence This thread concerns attempts to construct artificial general intelligence, which I often underline may likely be…jordanmicahbennett.github.io ▼ ▼ Author: I am a casual body builder, and software engineer.
[Article.Ai] Freaky — why Yoshua Bengio’s new “Consciousness Prior” paper is freaky to me
0
freaky-why-yoshua-bengios-new-consciousness-prior-paper-is-freaky-to-me-133d4382b6d5
2018-08-30
2018-08-30 18:44:55
https://medium.com/s/story/freaky-why-yoshua-bengios-new-consciousness-prior-paper-is-freaky-to-me-133d4382b6d5
false
409
null
null
null
null
null
null
null
null
null
Machine Learning
machine-learning
Machine Learning
51,320
Jordan Micah Bennett
Supersymmetric Artificial Neural Network: https://github.com/JordanMicahBennett/Supersymmetric-artificial-neural-network
4036d9858018
jordanmicahbennett
14
22
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-09-01
2017-09-01 17:18:43
2017-09-01
2017-09-01 17:23:06
1
false
en
2017-09-01
2017-09-01 17:23:36
2
133ecaabda6d
4.022642
0
0
0
Everyday I try to think of 10 ideas. Sometimes interesting, sometimes thought provoking, and often enough pretty stupid. The point is to…
4
10 Random Ideas for August 26, 2017 Everyday I try to think of 10 ideas. Sometimes interesting, sometimes thought provoking, and often enough pretty stupid. The point is to get the brain thinking, to exercise the brain muscle. None of these have been researched. All of them are original, as in I thought of them, but many may not be novel, as you know what they say, geniuses think alike :) Since ideas are useless if you don’t share them with the world, I decided to post them here. If you like any of them, go for it, make them a reality. If you ever want to chat about any of them with me, just look me up on sdbr.net or read my latest ideas on datarig.com. Thanks for reading! Here are my ideas for August 26, 2017: A book / movie about a world where people live in the Matrix like VR environment. The average human lifespan is still roughly 200 years, but to prolong their life, many people resort to “time slicing.” Time slicing is a technique where an AI controls an avatar 90% of the time, and the human is awakened only once a day or so to sync the memories and to be consulted on inputs. As a result, from a human point of view, they make all the decisions but can live 10 times longer. However, the story is told from the point of view of the AI, and how it deals with not really being in charge of it’s own destiny. First Steps: write a treatment for the main character and an outline. A service that takes a URL and converts it into a iOS and Android app wrapper code with push notifications capability managed by the company. This allows anyone to quickly convert an existing mobile website into an app. Make money from Push notifications and Store insertion fees. First Steps:I looked for this, but didn’t find anything that worked really well, or even marginally well. Do more research, if nothing, see how hard it would be to develop a basic wrapper with push on iOS. Wall clock with either an E-Inc or OLED clock face (e-inc would be more efficient). The clock connects to WiFi and pulls appointments from your calendar, and shows them on the clock as icons or free time. Different visualizations are possible. First Steps: check if it exists, if not, make a digital mockup and test market it. Service that allows operators of older, perhaps now outdated cameras to convert them to smart cameras with real time AI monitoring and notification. Box takes in older analog signals, encodes them, and sends them to a central server where AI monitors the feeds for unusual movements. Notifications with video clips are sent directly to guards via smartphone app. First Steps: check if it exists, talk to security companies to gather opinions. Running shoes with replaceable outer treads. The treads can be attached with tongue and groove connector similar to the zip lock bag, and a intertwine connector like on the lock type velcro. Benefits are that you can replace the treads every 200 miles without having to replace the entire shoe, which is eco friendly, and nice for your foot. First Steps: check if it exists and if it’s patentable. Pitch to Nike. Save anything public database backend. Are you writing a webapp that is mainly JS, and the only server backend you need is to save some user settings and data? This would be the service for you. The database would provide a RESTFul API to pull and push values and handle simple authentication. First Steps: research similar solutions, make a demo, test market it. Book about a college age girl who while interning in a physics lab accidentally triggers a door to a parallel reality universe, and falls through it. The parallel universe, by a happy coincidence, has the capability to sustain human life, but is different in every other conceivable way. Basically this is an update to Alice’s Adventures in Wonderland. This world is occupied mainly by robots and AI altered intelligent animals. The backstory is that AI came to be when humans of this universe developed it, but then humans became extinct due to an event that no one wants to talk about. The robots lost their purpose in life, to serve humanity, and they collectively decided to create biomechanical intelligence augmentation devices for remaining biological creatures on the planet. Of course, the various creatures, immediately after becoming more intelligent and learning how to talk, declare war on each other. Main character drops into the very middle of this war, and struggles to return home. First Steps:write character and world treatment, write an outline. Command line interface to everything. Write a tool for the Mac / PC that allows you to create command line driven macros for apps like Photoshop, Word, etc. The tool would present a command line where you can enter a string of macros, and they would be executed in the program of your choice. First Steps: check if exists, probably doesn’t. Figure out how to hook into functions of apps, or interact in different way. Applescript is nice, but not supported by all the apps. 2 Stage belt final drive for motorcycles that eliminates belt stretching by aligning secondary output with the rear swingarm pivot point. On many motorcycles the transmission output sprocket is slightly forward of the swingarm pivot point. By installing a through shaft through the middle of the pivot point and adding 2 sprockets on either side the tension problem can be solved. First Steps:check if patentable (prior art), probably, but why no one uses it nowadays? Come Back Later plugin for Chrome. A plugin that takes a screenshot of the currently opened tab and stores it and the URL into a list of websites that you don’t quite want to bookmark, but you have a feeling that you might need to come back to. First Steps: check if it exists, then write this if not. I need this :)
10 Random Ideas for August 26, 2017
0
10-random-ideas-for-august-26-2017-133ecaabda6d
2018-04-28
2018-04-28 01:29:06
https://medium.com/s/story/10-random-ideas-for-august-26-2017-133ecaabda6d
false
1,013
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Bogdan
I am a prolific creator, curious hacker, dyslexic writer, licensed and yet apathetic attorney, but above all else, I am human, and these are my ideas.
602ead4e1b00
x0054
57
3
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-08-30
2018-08-30 00:37:32
2018-08-30
2018-08-30 01:42:40
1
true
en
2018-08-30
2018-08-30 01:42:40
0
13403a9f9933
1.022642
0
0
0
Man, Who Got All His News From Reading Medium.com, Thought For Sure the World Would be Run By Well Learned Machines at This Point
5
Local Man From Last Year Surprised To Find His Pizza Not Delivered By Artificial Intelligence Powered Drone Man, Who Got All His News From Reading Medium.com, Thought For Sure the World Would be Run By Well Learned Machines at This Point Local Dominos artisan Bill Sapowski. He isn’t just a delivery guy. Image courtesy of PIxabay aka The Bay of Pixs. Local man Todd Stevens was shocked and surprised when the large pepperoni pizza he had ordered from the local Dominos was delivered to his door by a normal human driving a beat up Ford Taurus, and not dropped on his stoop by a gleaming artificial intelligence powered drone. “I could have sworn last year they said drones were going to be delivering food at the very least, by now. And I know I read that artificial intelligence was on the brink of complete control of the entire planet’s computer systems. Unless that dude who delivered my pizza was a bio-organic replicant synthesized to look exactly like high school dropout Bill Sapowski, there was no artificial intelligence involved in the delivery. I am also fairly certain his beat up Ford Taurus was not a quantum computer generated holographic representation of an autonomous robo-drone. Damn, that’s a disappointment. It does appear that America did finally collapse however, Dominos is still in business.” Todd then stuffed an entire slice of pizza in his mouth and began to cry.
Local Man From Last Year Surprised To Find His Pizza Not Delivered By Artificial Intelligence…
0
local-man-from-last-year-surprised-to-find-his-pizza-not-delivered-by-artificial-intelligence-13403a9f9933
2018-08-30
2018-08-30 11:24:32
https://medium.com/s/story/local-man-from-last-year-surprised-to-find-his-pizza-not-delivered-by-artificial-intelligence-13403a9f9933
false
218
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Daniel DeMarco
Research scientist (Ph.D. micro/mol biology), Food safety/micro expert, Thought middle manager, Everyday junglist, Selecta, Boulderer, Cat lover, Fish hater
7db31d7ad975
dema300w
3,629
148
20,181,104
null
null
null
null
null
null
0
null
0
19e43a17fcba
2017-11-21
2017-11-21 13:15:43
2017-11-21
2017-11-21 13:24:13
1
false
en
2018-01-15
2018-01-15 10:46:09
0
13405e1c2f3c
2.573585
11
0
0
Lots of projects enter the ICO. Some of them collect millions for their development. However, startups are more and more often given the…
5
AdHive has Successfully Passed the Reliability Test by Alexey Krol Lots of projects enter the ICO. Some of them collect millions for their development. However, startups are more and more often given the “scam” label in this environment. The reason lies within problems experienced by the team after successful fundraising or in the process of creating the product itself. For example, Tezos, which did nothing for several months at all. Due to the fact that no major blockchain project has made it to the final stage of development yet and because of the lack of the ICO market regulation, a cautious attitude towards the crypto industry can be witnessed more and more often. A crypto-enthusiast, blogger and expert in the Fintech industry, Alexey Krol offers his very own recipe for scam defining. He singled out a formula that helps to determine whether a project is operational or not. It is important for the AdHive team to meet all the criteria of a reliable startup, to make a statement that we are on the right track. You can read further these criteria. What is the problem of startups? All the blockchain projects that are going to enter the ICO stage today offer their fundamentally new solutions for problem solving in any direction. However, it is very difficult to assume how the development will actually proceed. After all, the team in fact solves the problem in a completely new way which has never been used before. Any new technology always faces some difficulties in the process of its creation. That’s why the team has to spend more time and money on it. This is the way not only the developers, but also all the scientists do work. How else can you create something new? Despite this, blockchain projects are very popular today. Their popularity is based not on the availability of opportunities only to participate in a dynamically growing market, but also on the fact that after the release of the product, the token holders are the first to gain access to the platform functionality. Alexey Krol notes that a serious project should meet the following criteria when entering the ICO stage: The project must have public importance: AdHive alongside with the public significance has an opportunity to initiate a market shift. Our platform is demanded by the public because it allows to optimize the advertising placement for one of the most effective types of channels — native videos; The problem solved by the team must be obvious and be explained in no more than 10 lines of the text: explanation of our project needs even less. We create a unique marketing platform for conducting an advertising campaign through bloggers. With its help, advertisers can post native video ad through bloggers in just 5 minutes, capturing by this a million people audience: brevity is the soul of wit; There should not be similar solutions within the variety of applied technological approaches in the world: we make the AdHive platform based on the synergy of two advanced technologies — AI and blockchain, Our platform has no analogues in the world! Serguei Popov, a well-known crypto-enthusiast in blockchain circles, an expert and co-founder of IOTA is an advisor of the AdHive project. That is one more proof of our prospect. He is a scientific consultant of our team, so we are confident in our model. And Alexey Krol’s test only confirmed the reliability of our developments. Before cooperating with us, Serguei Popov worked on the development of the WINGS mathematical model and conducted theoretical work on the NXT cryptocurrency. In AdHive, his model will allow the platform to distribute awards among users. Meeting all three criteria of Aleksey Krol, AdHive proves its prospects and solvency as a blockchain project.
AdHive has Successfully Passed the Reliability Test by Alexey Krol
444
adhive-has-successfully-passed-the-reliability-test-by-alexey-krol-13405e1c2f3c
2018-05-26
2018-05-26 05:25:31
https://medium.com/s/story/adhive-has-successfully-passed-the-reliability-test-by-alexey-krol-13405e1c2f3c
false
629
A community powered global network for native video advertisement.The AdHive platform automates all steps of interactions between bloggers and advertisers.AI modules for video and speech recognition connect to vlog channel and control the execution of the ad task by the blogger.
null
adhivetv
null
Adhive.tv
Adhiveinfo@gmail.com
adhive
AI,ADVERTISING,NATIVE ADVERTISING,INFLUENCER MARKETING
AdhiveTv
Blockchain
blockchain
Blockchain
265,164
AdHive.tv
The first AI-controlled influencer marketing platform on Blockchain. Launching massive advertising campaigns has never been so simple.
295e61003285
AdHiveTV
499
4
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-23
2018-07-23 14:57:01
2018-07-23
2018-07-23 15:15:26
1
true
en
2018-07-23
2018-07-23 15:15:26
1
13423a294160
4.381132
15
0
0
(this blog was originally written for the European AI Alliance; feel free to join this platform dedicated to European citizens for…
4
What we talk about when we talk about Artificial Intelligence (this blog was originally written for the European AI Alliance; feel free to join this platform dedicated to European citizens for discussing EU policies on AI) Nowadays, Artificial Intelligence (AI) is ubiquitous. We can hardly open a newspaper or tune in to a news show without getting some story about AI. AI is probably the technology most talked about. But AI means different things to different people. I’ve been working on the field of AI, both in industry as in academia since the late 80’s. Developed my first AI system in 86, an expert system to determine eligibility to social housing. Since then I’ve witnessed the deeps and the ups, the winters and the hypes in the field. Never before there has been this level of excitement, and fear, by so many, in so many areas, as we are seeing in the last couple of years. AI is breaking through in many different application domains, with results that impress even the most knowledgeable experts. Three main factors are leading this development: the increasing availability of large amounts of data, improved algorithms and substantial computational power. However, of these three only algorithms can be rightfully seen as a contribution from the AI field. More recently, awareness that AI has the potential to impact our lives and our world has no other technology has done before, is rightfully raising many questions concerning its ethical, legal, societal and economical effects. Government, enterprises and social organisations alike are coming forward with proposals and declarations of their commitment to an accountable, responsible, transparent approach to AI, where human values and ethical principles are leading. This is a very needed development, one to which I’ve dedicated my research in the last few years. Responsibility in AI begins with a proper AI narrative, which demystifies the possibilities and the processes of AI technologies and enables that all are able to participate in the discussion on the role of AI in society. In this short piece, I will try to explain what AI is, starting by describing what it is not. AI is not the Algorithm The “algorithm” is achieving magical proportions, used right and left to signify many things, de facto embodying, or seen as a synonym to, the whole of AI. AI has been around for give or take some 80 years, but algorithms are way older than that[1]. AI uses algorithms, but then so does any other computer program or engineering process. Algorithms are far from magic. In fact, the easiest definition of algorithm is that of a recipe, a set of precise rules to achieve a certain result. Every time, you add two numbers, you are using an algorithm, as well as when you are baking an apple pie. And, by itself a recipe has never turned into an apple pie. The end result of your pie has more to do with your baking skills and choice of ingredients. The same applies to AI algorithms: for a large part the result depends on its input data, and the ability of those that trained it. And, as we have the choice to use organic apples to make our pie, in AI we also have the choice to use data that respects and ensures fairness, privacy, transparency and all other values we hold dear. This is what Responsible AI is about, and includes demanding the same requirements from the ones that develop the systems that affect us. AI is not Machine Learning Machine Learning, and in particular Neural Networks, or Deep Learning, is a subset of AI that uses statistical techniques to enable computers to perceivesome characteristics of their environment. Current techniques are particularly efficient in perceiving images, and written or spoken text. By analysing many thousands of examples (typically a few million), the system is able to identity commonalities in these examples, which then enable it to interpret data that it has never seen before, which is often referred to as prediction. Even though, the results of current machine learning algorithms are impressive and go well beyond expectations, also this process is far from magic, but the result of applying well-known mathematical and statistical methods. Moreover, current algorithms are brittle (changing only one pixel in a picture can cause completely different classification), not easily generalisable (training has to start from scratch every time the machine should learn another task), and perception is just one component of intelligence, more is needed to achieve intelligent machines. A short definition of AI AI includes Machine Learning and is based on algorithms. However, the ultimate goal of AI is to develop computer systems that are able to simulate human-like intelligence. The term Artificial Intelligence was coined in the 50’s by John McCarthy, who defined it as the endeavour to develop a machine that could reason like a human, was capable of abstract thought, problem-solving and self-improvement. The challenge proved much harder than what those original scientists expected, and even current success of AI, in the area of Machine Learning, are very far from realising those objectives. The aim of this piece is not to speculate on the feasibility, or not, of this endeavour, but to provide a concise, practical definition. More than perception, AI is about reasoning. Besides machine learning, AI includes knowledge representation, planning, dealing with uncertainty, theorem proving, cognitive robotics and human-agent/robot interaction just to mention a few on the fields. Borrowing from the definition given in the seminal textbook on AI[2], I would say that AI is the discipline of developing computer systems that are able of perceiving its environment, and to deliberate how to best act on it in order to achieve its own goals, assuming that the environment contains other agents similar to itself. As such, AI is about autonomy to decide on how to act, adaptability to learn from the changes affected in the environment, and inter-actability to be sensitive to the actions and aims of other agents in that environment, and decide when to cooperate or to compete. A responsible, ethical, approach to AI will ensure transparency about how adaptation is done, responsibility on the level of automation on which the system is able to reason, and accountability about the principles that lead its interactions with others, most importantly with people. Notes [1] The word algorithm derives from al-Ḵwārizmī ‘the man of Ḵwārizm’ (now Khiva), the name given to the 9th-century mathematician Abū Ja‘far Muhammad ibn Mūsa, author of widely translated works on algebra and arithmetic. (source Wikipedia) [2] Russell and Norvig (2009): Artificial Intelligence: A Modern Approach, 3rd edition. Pearson Education.
What we talk about when we talk about Artificial Intelligence
69
what-we-talk-about-when-we-talk-about-artificial-intelligence-13423a294160
2018-07-23
2018-07-23 22:23:11
https://medium.com/s/story/what-we-talk-about-when-we-talk-about-artificial-intelligence-13423a294160
false
1,108
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Virginia Dignum
null
fb01d0a3bc3f
virginiadignum
72
11
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-05-06
2018-05-06 13:41:44
2018-05-06
2018-05-06 14:31:01
1
false
en
2018-05-06
2018-05-06 14:31:01
1
1342736e62d9
2.622642
5
0
0
Have you had the feeling that you want to get into Data Science, but were held back by the thought that it might be too complicated or…
3
Data Science..All Flavors Have you had the feeling that you want to get into Data Science, but were held back by the thought that it might be too complicated or “just for a math whiz?” Well, think again, because there’s never been a time where learning Data Science was this easy. With the abundance of online courses, blog posts and tutorials, you can tap into the basics of the field. Joining a community that helps support your learning process, meeting people in the field and exchanging experiences will boost your skills to the max. I want to talk today about my experience attending Dubai Data Science Meetup, the largest Data Science community in Dubai, UAE. I had the honor of being a speaker there as well. Writing about my experience will be from both viewpoints, as an attendee and as a speaker. The meetup was held on 5th May, 2018 at Microsoft Office in Dubai Internet City. It was attended by more than 100 people, different levels of experience, but all sharing the same professionalism and enthusiasm. Link to the meetup page: https://www.meetup.com/Dubai-Data-Science-Meetup/events/248867575/ The first speaker was Aishwarya Ramachandran, Data Analyst at Microsoft. Her topic was “Demystifying Machine Learning.” She started the talk defining what is Machine Learning and explaining its different types and best use cases for each. Then she showed a demo model for predicting used car prices, using Microsoft Azure ML Studio. The talk was targeting beginner level, where people could quickly grasp the basics of ML and how to simply build a model without coding, using automated tools and cloud platforms. The second speaker was myself, an IBM Cloud Developer Advocate. I took the experience to the next level, where I targeted intermediate skill levels who are more comfortable working with code, python in specific. I discussed how to build a “Customer Churn Prediction Model.” I showed the whole pipeline in a Jupyter notebook. Starting from reading the data, to cleaning and manipulating it, to preprocessing, selecting and training some models to finally evaluating them to choose the best for production. As a bonus content, I showed how to deploy the selected model on IBM Cloud using Watson Machine Learning client library. I also showed how to do a REST API call with data sent to the model and getting the prediction back as JSON response. The third speaker was Vitalii Duk, Lead Data Scientist and Machine Learning Engineer. His talk complemented the previous speakers’ topics and leveled up the experience to the max. His talk was targeting advanced and seasoned Data Scientists. He discussed Machine Learning at scale using an open source python library created specifically to handle big data tasks and heavy workloads. This library is called “Dask.” Personally, I learned a lot from this talk and will definitely incorporate some of the ideas into my upcoming workflows. Lastly, the final speaker was Pavankumar Gurazada, who talked about R as a language of choice for Data Science and how C++ can be used in exchange with R to handle complicated tasks in a performance-efficient fashion. All the sessions had very good interaction from the audience, with interesting questions and interesting side discussions. The audience was very appreciative for the informative sessions and the mostly heard feedback was that they came out with a lot of value from this meetup. It was a great experience for me and I will definitely join next meetups. Being a self-learner and an online learning advocate, I found a new treasure, the power of communities. It’s added to my list as one of the best and most recommended ways to learn, meet field experts and enthusiasts through meetups, get to hear them talk about the latest technologies they use, ask them about their experiences and exchange field insights. There’s never been a better time for learning…
Data Science..All Flavors
62
data-science-all-flavors-1342736e62d9
2018-05-06
2018-05-06 16:41:37
https://medium.com/s/story/data-science-all-flavors-1342736e62d9
false
642
null
null
null
null
null
null
null
null
null
Data Science
data-science
Data Science
33,617
Heba El-Shimy
null
137044c5f74b
heba.n.elshimy
12
14
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-03-08
2018-03-08 21:17:19
2018-03-08
2018-03-08 21:17:16
1
false
en
2018-03-08
2018-03-08 21:17:20
9
1342b510fcca
5.184906
0
0
0
null
5
Salesforce Sales Cloud Supercharged — Einstein’s next Move The News A few days ago Salesforce announced an update to its sales cloud that features Einstein powered predictions, insights, and productivity. The press release is linked above or alternatively you can read it below, along with some comments of mine. Salesforce is (again) addressing the three main issues that plague CRM implementations since Tom Siebel coined this term. Let me paraphrase them> Salespeople do not find the time to do their job, which is selling. Instead they are spending an inordinate amount of time entering data that supposedly only helps their management controlling them a little more. Sales managers do not have enough visibility into what is going on in their area of responsibility, what their team is doing (and why), whether they are doing the right thing. The same problem, of course, applies to the Head of Sales, just at a bigger scale. Sales operations is charged with creating meaningful reports that tell the one single truth. This they need to do using data that resides somewhere, data that is distributed, instead of some central consolidated place. Data that is essentially not fully trustworthy. Salesforce is doing this using a triple of features: The Salesforce resident AI: Einstein to help sales persons identify the most promising opportunities to work upon The Salesforce Inbox that increases productivity by attributing emails to the right accounts as well as connecting to the calendar Sales Analytics to help salespeople and their management to visualize, interpret, and use the available data The Press Release Ask any rep what their favorite part of the day is, and chances are that their answers won’t involve the words “logging” or “data.” Ask managers what they really want from their sales teams, and I bet you that they’ll ask for more visibility into what their reps are doing and for their reps to spend more time talking to customers, building relationships. As for sales operations? They probably would prefer if answering business questions didn’t involve four systems, two excel sheets, and a pivot table. These traditional systems don’t set sales teams up for success today or in the future. In fact, having all of these disparate systems causes sales reps to spend 25% of their time logging data instead of doing what really matters — building relationships and selling. Which is why we’re introducing a supercharged Sales Cloud Einstein. Bringing together Sales Cloud Einstein, Salesforce Inbox, and Sales Analytics, to deliver more predictions, insights and productivity gains than ever before. It’s bringing the power of artificial intelligence to every step of the sales process. So how does it all work? Great question. Let’s break it down. AI with Sales Cloud Einstein AI is continuing to take center stage, revolutionizing the way we work. With Einstein, AI prioritizes focus on the most critical areas to help every sales rep increase their productivity and win rates. Features like Einstein Lead Scoring can turn mountains of data into critical signals that have the power of identifying the leads that are most likely to convert, and Einstein Opportunity Scoring can identify a poorly-performing opportunity proactively so a sales rep can keep it on track — before it falls off. Einstein Opportunity Insights give reps the ability to address at-risk deals and learn best practices from the most successful ones. And, when reps are armed with the right insights about their accounts’ businesses, conversations become more efficient and effective, enabling reps to sell more. Productivity with Salesforce Inbox Did you know that on average, sales reps spend 64% of their time on non-selling tasks like data entry? With Salesforce Inbox reps can maximize the time they spend selling by taking advantage of automated data capture — all of those emails are logged to the right records, automatically. With built-in email productivity, reps can eliminate the hassle of scheduling meetings. Plus, they can see their top email priorities, right on their phones, and, get visibility into their opportunities, leads, accounts, and contacts. Reports and Dashboards with Sales Analytics Instead of exporting, collating, aligning and analyzing, sales teams can just click in to built-in analytics, with ready-made dashboards that make it easy to understand and explore whitespace and team performance. Reps can uncover pipeline trends, and take action immediately. Sales leaders can understand how the team is functioning across regions and products, and identify top sellers, as well as those that may need more coaching. And, it’s easy to analyze deals from lead to close with over 40 out-of-the-box KPIs. By combining all three products, we’re able to create a Sales Cloud Einstein that is a predictive data scientist for every sales team. A constant companion that drives productivity through efficiency and insights. No more point systems, no more multiple contracts, no more copy-and-paste, no more switching. Just a clean, easy, modern solution to make every company a smarter, more efficient, more productive, customer-focused company. Sales Cloud Einstein (now including Salesforce Inbox and Sales Analytics) is priced at $50 per user, per month. Click here to learn more about Sales Cloud Einstein. The Bigger Picture Artificial intelligence / machine learning will continue to permeate business applications. The good news is that there is less and less talk about AI, if not in a kind of personal way, like Einstein or Leonardo (and yeah, I know that Leonardo does not only cover AI). This shows that the big vendors are more and more going away from a technology narrative to a results narrative. AI is not a means by itself but a means to an end. And this end is doing more with less effort. My PoV and Analysis By addressing these three pain points of sales teams Salesforce is partly using some pages out of the Nimble playbook. Especially the mobile app for the Salesforce Inbox looks remarkably familiar to me. I do not say that this is a copy, but that moves like this one are actually inevitable, and that CRM (and CEM) solutions will become ever more similar. The combination of these three features and how they are laid out now tells the same story that I heard about the same time from clari, a company founded in 2013 and that focuses on “transforming the way they [companies] sell, make decisions, and grow”. Now, SAP, Oracle and Microsoft are telling the same story. This, along with the observation that even small companies are more and more referring to the term platform, reiterates that business applications are commoditizing fast and that the main battleground is becoming the fabric of a business, i.e. becoming the platform of choice. For Salesforce this is an important move as, all Salesforce “hype” taken away, there is a perception out in the market that the Salesforce architecture is aging. On top of this Salesforce is not a bargain (judging by the list prices) nor is the company riding on a high profitability. And the other 3 of the big four can tell end-to-end stories that cover the full value chain. Plus all of the big four are clawing their way from the enterprise market down to the midmarket and eventually the smaller businesses, which is the place where the opportunity lies. With that, Microsoft, Oracle, and SAP being able to tell a story of a more modern platform, and being attacked by small and nimble players that can ride on Salesforce’s and other platforms, Salesforce needs to continue to show that it is on the forefront of innovation with the benefits of all user groups in mind. I think that this worked out fine with this release. Salesforce is at least on par with the competition. But is this a super charged sales cloud? No, it isn’t. Unless the previous engine was seriously underpowered — which it wasn’t. The race towards becoming the platform of choice continues.
Salesforce Sales Cloud Supercharged — Einstein’s next Move
0
salesforce-sales-cloud-supercharged-einsteins-next-move-1342b510fcca
2018-03-08
2018-03-08 21:17:21
https://medium.com/s/story/salesforce-sales-cloud-supercharged-einsteins-next-move-1342b510fcca
false
1,321
null
null
null
null
null
null
null
null
null
Einstein
einstein
Einstein
635
Thomas Wieberneit
Helping businesses to improve in Digital Transformation, Customer Engagement, Customer Experience, CRM, Innovation
1968374dc9da
twieberneit
1,370
1,311
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-09-29
2017-09-29 13:00:33
2017-09-29
2017-09-29 13:32:34
1
false
en
2017-09-29
2017-09-29 13:32:34
5
13439ee0cbaf
4.569811
2
0
0
Open on a busy downtown street–it’s evening, cold and drizzling. Our traveler walks into a lush, well-lit hotel lobby after an arduous…
5
Empathy for the Machine — Design Considerations for Voice Photo by Drew Coffman on Unsplash Open on a busy downtown street–it’s evening, cold and drizzling. Our traveler walks into a lush, well-lit hotel lobby after an arduous trip. The modern chic decor is stunning. No doubt the service here will be excellent, she imagines, as her thoughts drift to imminent room service and a plush comforter. She walks up to the desk to check in and sees no one. Meanwhile, our zero UI auto-concierge waits with bated breath under the front desk: [urgent whispering] “I’m down here–just say the word! I know you! Your previous rooms, your meal preferences, and even that it’s your birthday tomorrow–I have a surprise for you! Just notice me…say the words to invoke my check-in application and I’ll be at your service!” The weary hotel guest raps her nails on the marble, looks around, rings the bell…her gaze eventually falls to a small branded placard with a bulleted list of written instructions for screenless check-in. She rings the bell again… Well, that sounds like a modern service disaster of Richard Scarry proportions–the kind where all the trucks on the road collide and there is mustard and maple syrup and hot dogs everywhere. Yes, I am aware that it is unlikely that a prestigious hotel chain would consider replacing a crucially important component in the service chain with anything like a basic Alexa Skill. Regardless of technology, there are sticky design problems that need to be effectively addressed if we want to reach the goal of zero UI bliss — technology that responds to people in ways that feel as if it were alive. THE DESIGN PROBLEMS REMAIN THE SAME Obviously, a voice interface may not be the best road to excellent service for this particular hotel service application. Even with a custom voice application, some likely hurdles for this scenario might be: How do we make the user aware that voice check-in is possible, and encouraged? How do we help our traveler get started? What does she need to say? How do we build trust and deliver accurate, top-notch service? How do we address all of the potential questions or problems our traveler (or any traveler, actually) might have? Can we do this? How do we sense frustration and turn the interaction around? From a human-centered design perspective, we would start from the perspective of the traveler. We would empathize, explore and prioritize different possibilities in the traveler’s journey, then iterate and solve the problems. However, screenless solutions excel and come to life when we can “humanize” the interaction, meeting users on their terms. To address these additional complexities, I’m proposing an additional layer of consideration to help put an interaction in perspective: empathize with the machine. Treat the machine like the other side of a conversational interaction. Looking at the problem from both the user’s and the machine’s perspectives can help you better understand the opportunities and gaps. So, just how might we go about empathizing with the machine? Assuming you have identified a problem worth solving for users and your business, and you believe a conversational interface might be a good avenue to explore, consider the following: PLAY TO VOICE INTERACTION STRENGTHS It’s not magic, it’s technology. Well executed, conversational interfaces can be a fantastic way to increase a feeling of intimacy, save time and physical effort, or serve situations where being tied to a screen is inconvenient or less than ideal. But some scenarios are more ideally suited for voice than others. Voice interfaces can be great for: Hands- or eyes-free environments where safety or screen visibility is a concern Complicated things that people can articulate (for example, “Show me psychological thrillers from the 1990’s, available for free, with a rating of 3 or more stars”) Languages that are hard to type On the other hand, voice interfaces are generally not good for: Input that is hard to describe Situations where negotiation or lots of variables are required Huge amounts of input or output adapted from Laura Klein’s Designing from Voice Interfaces Another consideration is that voice interfaces without screens often lack sufficient visual cues to indicate to users that 1) voice interaction is possible, and 2) here’s how you do it. Often, paper “tooltips” are used but then again, this is getting away from zero UI. Voice interaction standards have yet to solidify across devices, so there’s no clear and obvious way to start up an interaction, yet. With this in mind, is the scenario you’re considering suited for a voice application? View it from the machine’s perspective, playing to its strengths in order to set it up for success. RESEARCH, EXPLORE, AND TEST POTENTIAL INTERACTIONS. REPEAT. Human interaction is nonlinear by nature and the type of input to which a machine needs to respond can vary greatly. In order to make the machine side of a zero UI interaction feel lifelike and human, take the necessary time to research and understand the numerous ways users can provide input. Think through the scenarios, experiment, test with users, and iterate. Do this early and often. Also, be mindful of the scope of an interaction–how much are we asking of the machine and what does it need to do? As the potential inputs and types of tasks grow exponentially, does the overall experience begin to suffer? If it’s too broad, focus it. Consider the elements that make for a great core interaction and build from there. UNDERSTAND CONTEXT AND INTENT When people talk to one another, so much information is conveyed by nonverbal cues like tone, body language and facial expressions. Using natural language is key to making the interaction feel like a 1:1 conversation, but a surface level human-esque veneer can set up unrealistic expectations if the machine cannot demonstrate an understanding of situational context, build trust via accuracy and consistency, recognize return users, or retain and reference a history of interactions. When designing for a specific voice application, work to find out what the machine needs to know in order to respond effectively and be truly useful. What data can it access? What critical cues, based on content, inflection or tone could help bring the interaction to the desired level of “human-like” that makes the overall interaction a success? Of course, machines are still machines. Offer elegant ways out of a jam, providing an ability to sense a problem and elegantly recruit an actual human to intervene. KEEP EMPATHIZING It’s critical to develop empathy for users when practicing human-centered design, just as it has always been. But when crafting human-like zero UI interactions, strive to gain perspective from both sides, human and machine, in order to create more lifelike experiences. This post originally appeared on the PointSource Blog, where I work as a designer. Interested in chatting more about voice or zero UI design, or perhaps working with me? Great — let’s chat!
Empathy for the Machine — Design Considerations for Voice
11
empathy-for-the-machine-design-considerations-for-voice-13439ee0cbaf
2018-03-21
2018-03-21 06:36:53
https://medium.com/s/story/empathy-for-the-machine-design-considerations-for-voice-13439ee0cbaf
false
1,158
null
null
null
null
null
null
null
null
null
Zero Ui
zero-ui
Zero Ui
15
scottmccall
Designer @Globant, guitar twangler, synth mangler. I am fascinated by intuitive interaction, breathtaking sounds, and curries of all sorts.
70b4f9ed9213
scottmccall
153
239
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-05-18
2018-05-18 10:13:46
2018-05-18
2018-05-18 11:43:41
2
false
ru
2018-06-08
2018-06-08 19:36:00
6
1343c3ec6389
3.594654
11
0
0
Мы в HFLabs создаем робота, который приспосабливается к произвольной среде и каждый день может вести себя по-разному.
3
Автономный робот. Начало Мы в HFLabs создаем робота, который приспосабливается к произвольной среде и каждый день может вести себя по-разному. С таким роботом можно играть. Можно целенаправленно создавать для него игры, обучающие конкретным навыкам. А можно просто оставить робота наедине, чтобы учился сам. Гипотеза в том, что такая игрушка никогда не надоест, поскольку все время эволюционирует. Мы ориентируем ее на тех, кто не боится завести умную штуковину, у которой свои цели, неизвестные и разные. Целевая функция роботов и людей В классической робототехнике роботов стараются делать полезными для людей. Это значит, что есть целевая функция, которая оценивает полезность. «Быстрее», «точнее», «устойчивее», «надежнее» – это все целевые функции. И разработчики, и сам робот, если он с элементами самообучения, максимизируют целевую функцию действиями или бездействием. У классического подхода уже очень много достижений: беспилотные автомобили, распознавание образов и речи, игра в шахматы и го, подбор рекомендаций по истории покупок. А какая целевая функция у человека? На этот вопрос огромное количество ответов, от религиозных до гедонистических. Более того, у каждого на разных этапах жизни функция разная. Иногда – просто добежать до туалета, иногда – забабахать что-то надличностное, что переживет автора. Мы пришли к выводу, что любой агент, живой или искусственный, стремится к состояниям, из которых, по его мнению, доступно как можно больше других состояний. Человек старается избежать смерти, серьезных увечий и болезней, поскольку они резко сокращают доступные состояния. В случае смерти – до нуля. Но если веришь в бессмертную душу, появляются случаи, когда правильная смерть лучше неправильной жизни. Деньги расширяют пространство состояний, поэтому людям нравится, когда у них больше денег. Целевая функция человека и автономного робота – достичь состояния, из которого доступно как можно больше других состояний. Различимые и неразличимые состояния Состояние – это все ощущения, которые человек получает в каждый момент времени. Всё, что он видит, ощущает и думает. Казалось бы, люди все время получают что-то новенькое. О каком расширении или сужении пространства состояний можно говорить? Даже если сидеть неподвижно и смотреть на стену, скоро начнешь замечать ранее невидимые детали. Это и есть новые состояния. Если закрыть глаза, почувствуешь, как тело с тобой разговаривает. Начнешь чувствовать движение воздуха внутри носа. Все это очень даже новые состояния. Поэтому правильнее говорить о состояниях, которые различимы. Различимые состояния – те, которые ведут к разным результатам в контексте важных для конкретного человека целей. Сидеть на месте для того, кто ориентирован на зарабатывание денег, – бессмысленно и скучно. В это время одно состояние неотличимо от другого. Сидеть на месте для человека, практикующего медитацию, – увлекательнейшее путешествие в самого себя. Получается прекрасная система, определенная в терминах самой себя. То, что человек считает одинаковым или разным, зависит от его целей. А цели – это состояния, из которых доступно максимум других различимых состояний. Из такой системы следует, что каждый в среднем получает примерно то, во что верит. Однако поверить во что попало человек не может, потому что даже восприятие объектов и явлений одинаковыми или разными зависит от того, во что человек верил раньше и в каких состояниях оказывался. Человек различает только те состояния, которые ведут к разным результатам в контексте его целей. Стремление к новизне Если испытываемые состояния слишком одинаковы, обычно это означает, что человек никуда не движется. Он чувствует скуку, дискомфорт. Дискомфорт важен, потому что с застывшими формами безжалостно расправляется красавица-энтропия. Но если испытываемые состояния отличаются очень сильно от спрогнозированных и друг от друга, возникает паника потери контроля. В слишком меняющейся среде почти невозможно ничего прогнозировать, мозг становится попросту не нужен. Паника потери контроля – это когда водитель из Мурома получил права три дня назад и пытается проехать по Таганской площади в Москве. Человек стремится к новизне, но умеренной. Суть проекта автономного робота Всю эту философию мы хотим воплотить в физическом роботе. У него будет две задачи: расширять пространство состояний согласно своим целям: предположениям о тех состояниях, из которых доступно больше других состояний; удерживать долю получаемой новизны в определенном небольшом диапазоне. Цели робота будут по мере развития меняться, а мотивация на новизну – нет. В активной среде, он будет уделять больше внимания многократным повторениям, чтобы делать среду более предсказуемой. В пассивной будет чаще менять поведение, чтобы взаимодействовать со средой новыми и новыми способами. Глобальная цель – технология создания автономных штук, способных жить в любой среде. Им нельзя будет приказывать, только договариваться. Договариваться теми методами, которые им доступны и укладываются в их цели. А прямо сейчас у нас есть простой робот, сделанный из пылесоса. С камерой и колесиками. Мозг находится не в роботе, а в компьютере разработчика Пока мы работаем над тем, чтобы робот произвольно двигался, наблюдал за своими движениями и изображением на камере и периодически возвращался на зарядную станцию. Это единственное для него ограничение, а в остальном – свободная игра и полет. Если все это вам интересно, присоединяйтесь. Подписывайтесь на блог, приходите в команду. Мы открыли две вакансии: редактор блога (40 000 рублей на руки, частичная занятость); программист на python (150 000–200 000 рублей). Пока у нас только пылесос с камерой. Есть шанс войти в проект на старте. О чем мы расскажем дальше Почему люди решили, что боль – это плохо; почему на самом деле речь не про состояния как таковые, а про субъективно различимые состояния; граница между порядком и хаосом – почему в армии дисциплина, а в дизайнерских компаниях бардак. Подписывайтесь на обновления: Facebook; VK; Telegram.
Автономный робот. Начало
31
автономный-робот-начало-1343c3ec6389
2018-06-08
2018-06-08 19:36:01
https://medium.com/s/story/автономный-робот-начало-1343c3ec6389
false
851
null
null
null
null
null
null
null
null
null
Роботы
роботы
Роботы
145
It Just Lives
Создаем робота, который живет как хочет
8f28972d7f19
it.just.lives
7
2
20,181,104
null
null
null
null
null
null
0
null
0
c4d1023640a8
2018-01-18
2018-01-18 09:29:26
2018-01-19
2018-01-19 10:45:26
6
false
en
2018-01-19
2018-01-19 10:55:20
14
13442444bbdf
6.836792
13
0
0
Last year, we accidentally built our own chatbot framework. Three implementations later we’re still optimistic about enterprise automation.
4
A Year With Orbit Operator Last year, we accidentally built our own chatbot framework. Our colleagues at the Bakken & Bæck AI research lab in Bonn had already developed an AI service for understanding natural language in 2013/2014, before we got into writing robots in 2015/2016. It seemed like a good idea to put the two together by making a template for automated conversations: Orbit Operator. Arguably, we started this project at the wrong end − with the technology instead of the problem − but the real push to actually build Orbit Operator mostly came from our chatbot-curious clients in service intensive consumer industries. And so it came to pass that, by the end of 2017, three virtual assistants went live on Orbit Operator: An infobot for the City of Oslo. The bot provides authorized parenting advice, curated by an experienced medical staff. Parents of young children often find themselves searching for answers online, a quest that usually ends up in getting ‘hearsay’ advice on parent forums or Facebook groups. The City of Oslo bot is based on professional expertise and provides quick answers to new parents’ FAQs. The infobot aims to free up time for the medical staff, so that they can spend more of their precious minutes on personal, face-to-face conversations. Furthermore, it aims to extend the municipal health station’s opening hours without increasing the staff’s workload. A customer service assistant for media and classified ads giant Schibsted. The chatbot answers FAQs from subscribers to five Norwegian newspapers and it provides assistance by performing simple customer service tasks, like solving login issues, sharing digital subscriptions and changing the delivery address for the newspapers. The bot is integrated with Schibsted’s call center technology stack. A bot that collects customer reviews for restaurants. While dining out, customers can upload a picture of their dish (hey, if it’s good enough for Instagram) and give an instant thumbs up or down to their food experience. This bot was developed with HORECA.digital, German retailer METRO’s digital arm, a subsidiary that equips independent restaurants, hotels and caterers with innovative digital tools. Hold that line. Another chatbot platform? With so many chatbot platforms out there, the legitimate question ‘why in the world should we make our own?’ naturally popped up. The overwhelming number of chatbot frameworks that have emerged since bots became the new apps (thanks again, Microsoft!) provides a partial answer to this question: it is extremely difficult to come up with one generic recipe for building full-service chatbots that meet real user needs. A ‘useful’ chatbot does at least three things well: It understands what the user wants by identifying questions, greetings, idle chit-chat, requests and answers. It thinks and makes decisions by processing the input from its users, asking them for clarification, searching through databases and checking business rules. It responds and acts by composing a reply, presenting a suitable predefined answer or performing tasks on behalf of the user (e.g. ordering a meal, changing the address for a delivery, etc.) “Understanding” has become easier for bots thanks to commoditized, open source technology for natural language processing (e.g. Google Cloud Natural Language API). Once the bot understands, its “responding” mostly boils down to presenting a pre-defined text (similar to the standard answers one gets from a junior customer service representative). The “thinking” part is more tricky. Finding a generic solution for chatbot “decision-making” is significantly more difficult than writing a recipe for understanding human language. The act of ordering an Uber is pretty self-explanatory, but the right way to handle a complaint very much depends on the organization that provides the particular service. An appropriate response for one company, could be an inappropriate response in a different company context. Standardizing the “decision-making” part in a chatbot framework is nearly impossible because the inner workings of organizations (even in the same industry) are all slightly different. Training a chatbot to perform organization specific tasks, means diving into a whirlpool of business rules and legacy systems. This is one of the problems we chose to tackle by building Orbit Operator as a framework. This way, we could make our developers’ lives a bit easier, while at the same time providing different kinds of organizations the tools to customize the bots to their needs. Half-Baked, not Half-Assed. What is Orbit Operator and what does it do? Orbit Operator basically provides a chatbot development recipe on how to mix open source algorithms for natural language processing with custom integrations to enterprise systems. This combination results in a particular dish of averagely intelligent digital assistants, which have a sound understanding of user queries and are above-average efficient in solving them. Orbit Operator consists of: A chat widget. Oftentimes a standard messaging app user interface like Facebook Messenger or WhatsApp will do. Sometimes, however, organizations prefer to embed a chat interface into their own communication channels (e.g. native app or customer account pages). Operator works for both scenarios, but a custom chat interface makes it easier to accommodate particular requirements e.g. data privacy concerns or branding/styling. Pre-trained machine-learning models for understanding user intents, which include text classification with language support for English, Norwegian and German and image recognition. An admin panel that lets enterprise users train the chatbot and set up dialogue templates, in order to effectively pass on their expertise (and boring tasks) to a digital assistant. Conversation flow logic that can be customized with customer specific elements like predefined responses, steps for process automation and business rules. Customizing dialogue flows with Orbit Operator Human handover. We all know how frustrating customer service conversations can be when the bot doesn’t get it right the first, second or eleventh time. So we made it easy for users to speak to a human being in case they need to. Hooks for integration with customer care technology stacks, e.g. Schibsted’s customer service bot that is integrated with Siebel CRM and the customer’s existing call center solution Puzzel. Customer insight gathered from Chatbase analytics. The five most important lessons Operator taught us Building digital assistants for media, retail and the government sector taught us a thing or two (or five) about chatbots: 1. ‘I want to speak to a chatbot’, said no one ever. Chatbots are first and foremost a productivity tool. While it’s true that customers − especially the younger ones − would rather text than talk, the real (and most valid) reason to invest in enterprise digital assistants for process automation is to increase cost efficiency. Being honest about this motive (and future perspective) makes it easier for customer service teams to adopt a chatbot as an important software tool in their work routines. 2. Training data is the gift that keeps on giving — but only if you give it enough time. Launching a chatbot is just the start of the journey. To reap the rewards of automation, organizations need to have a team/role in charge of monitoring bot responses and correcting bot behaviour by adding training data where needed. 3. Content is king, also for chatbots. We spent about one third of our time on defining the right content, in the form of both training data and predefined chatbot responses. This time investment was well worth it, first of all in terms of better-trained and more precise algorithms, but also in giving the organizations we work with an insight in how chatbots actually work. Involving the ‘chatbot trainers’ early on in testing the NLP models definitely paid off, as did visualising the test results. By asking experienced medical staff to curate the answers provided by our Oslo health infobot, we made sure that they were involved to such an extent that they actually cared enough to make it better. Visualizing test results for the NLP model 4. Remember to check the right language box. At some point after launch, we noticed that the City of Oslo’s infobot was unusually hung up on ‘breastfeeding’. Granted, this is a topic of high interest among our users, it was overrepresented in our training data set as well, but still something seemed off. As it turned out, we forgot to check the language setting — the bot used English rules to understand Norwegian. 5. You need a panic button. There is no doubt that today’s technology for conversational agents has limitations. To acknowledge and prevent user frustration with chatbots ‘under training’, we designed a chat experience with a visible ‘take me to a human’- panic-button. The by-product of this feature is a way to measure the relative degree of automation, by comparing the number of customer requests fully solved by the chatbot with the number of cases that are handed over to a human customer service representative. If you can measure it, you can improve it! There is a tendency among impatient techno-optimists to declare chatbots dead, even before they are fully-grown. Chatbot technology shortcomings are evident in the demise of all-purpose personal virtual assistants, like for instance Facebook M. We, however, like to think of ourselves as techno-realists, as we believe in the combination of AI and human capabilities –as already attempted by virtual assistants like Fin. While the current status of enterprise chatbot technology is far from flawless, we do believe in its potential and therefore we’re not planning to give up on Orbit Operator any time soon. Viva La AIssistance!
A Year With Orbit Operator
155
a-year-with-orbit-operator-13442444bbdf
2018-05-30
2018-05-30 15:05:10
https://medium.com/s/story/a-year-with-orbit-operator-13442444bbdf
false
1,560
The Bakken & Bæck blog
null
bakkenbaeck
null
B&B Stories
null
bakken-bæck
BAKKENBAECK,DIGITAL AGENCY,TECHNOLOGY,DESIGN
bakkenbaeck
Technology
technology
Technology
166,125
Codruta Gamulea Berg
null
73d328ab82dc
codrutagamulea
55
62
20,181,104
null
null
null
null
null
null
0
null
0
60b40d78a847
2018-07-31
2018-07-31 11:27:47
2018-07-31
2018-07-31 13:40:40
1
false
es
2018-08-27
2018-08-27 11:00:04
4
1345deef0903
2.14717
1
0
0
“Aprender sin reflexionar es malgastar la energía.”- Confucio
5
APRENDE PYTHON YA! Libro: Los pilares de Python. #3 - Breve historia y características “Aprender sin reflexionar es malgastar la energía.”- Confucio A pesar de que el nombre de Python nos recuerde al conocido reptil, el nombre que eligió su fundador, Guido van Rossum, en 1991 fue como consecuencia de uno de sus shows preferidos de la BBC, Monty Python. Actualmente es utilizado por miles de compañías y millones de programadores principalmente por las siguientes características: 1- Es un lenguaje gratuito, fue creado bajo la licencia de software libre. Es decir, que cualquier persona puede distribuir copias del software, realizar cambios o leer su código fuente. 2 - Es un lenguaje interpretado: Hay lenguajes de programación como C o C++ que son convertidos a binario (0 y 1) utilizando lo que se conoce como un compilador. Sin embargo, Python no necesita dicho compilador y convierte el código fuente en un código intermedio llamado bytecode. Este “detalle” significa un ahorro de tiempo a la hora de evitar tener que compilar el programa y asegurarnos que las librerías están correctamente cargadas con nuestro compilador. 3 - Es un lenguaje multiparadigma, es decir que para afrontar un mismo problema puedes utilizar varios “enfoques” o “estilos de programación”. 4 - Integrable: Como si de una pieza de Lego se tratara puede unir partes escritas en Python con otros lenguajes de programación como C++ y así obtener lo mejor de los mundos. 5 - La gran comunidad (personas que programan en el mismo lenguaje) hace que cualquier problema al que te puedas enfrentar haya sido probablemente ya resuelto por otra persona evitándote más de un dolor de cabeza. Como consecuencia directa hay un gran número de librerías o programas creados por otros usuarios con el objetivo de simplificar alguna tarea. 6 - Sintaxis simple: Como ya habíamos visto en el anterior capítulo, la sintaxis de Python es probablemente una de las más simples y fáciles de entender todos los lenguajes de programación aumentado la agilidad para leer y escribir código. “An overhead shot of a blank notebook, a magnifying glass, a camera and several other exploring-related items” by Dariusz Sankowski on Unsplash Finalmente, antes de empezar a escribir código, hay que advertir que actualmente python se encuentra en un momento de transición. Al igual que paso en el año 2000 cuando los países europeos circulaban las monedas euro y locales a la vez, en python convivimos con la versión de Python 2 y su equivalente más moderno Python 3. A finales de 2020 Python dejará de dar soporte la versión Python 2. Lógicamente, en este libro trabajaremos únicamente sobre la versión de python 3. — — — — — — — — — — — — — — The End — — — — — — — — — — — — — Si te gusta esta pequeña y gratuita revista puedes ayudar simplemente compartiéndola o suscribiéndote a la revista. Soy Rubén Ruiz, trabajo en la industria financiera en Inteligencia Artificial. Como proyecto personal llevo esta pequeña revista donde experimentamos con Inteligencia Artificial… Puedes también seguirme en: Instagram (Vida personal, es divertido) => @rubenruiz_t Youtube (Canal sobre IA, intento que sea divertido )=> Rubén Ruiz A.I Github (Donde subo mi código, esto ya no es tan divertido) => RubenRuizT
APRENDE PYTHON YA! Libro: Los pilares de Python. #3 - Breve historia y características
1
aprende-python-ya-libro-los-pilares-de-python-3-breve-historia-y-características-1345deef0903
2018-08-27
2018-08-27 11:00:04
https://medium.com/s/story/aprende-python-ya-libro-los-pilares-de-python-3-breve-historia-y-características-1345deef0903
false
516
Experimentos con Inteligencia Artificial.
null
null
null
AI experiments en Español
rubenruiz90@gmail.com
ai-experiments-en-español
PYTHON,PROGRAMMING,ARTIFICIAL INTELLIGENCE,R,DEEP LEARNING
null
Python
python
Python
20,142
Ruben Ruiz
null
2db774b0464f
rubenruiz_26771
24
22
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-03-15
2018-03-15 17:06:42
2018-03-15
2018-03-15 17:08:29
1
false
en
2018-03-15
2018-03-15 17:10:38
53
1348f5eb14fe
7.249057
1
0
0
Standard software development practices for web, Saas, and industrial environments tend to focus on maintainability, code quality…
3
Jupyter, Zeppelin, Beaker: The Rise of the Notebooks | Open Data Science Standard software development practices for web, Saas, and industrial environments tend to focus on maintainability, code quality, robustness, and performance. Scientific programing in data science is more concerned with exploration, experimentation, making demos, collaborating, and sharing results. It is this very need for experiments, explorations, and collaborations that is addressed by notebooks for scientific computing. Notebooks are collaborative web-based environments for data exploration and visualization — the perfect toolbox for data science. In your favorite browser, you can run code, create figures, explain your thought process and publish your results. Notebooks help create reproducible, shareable, collaborative computational narratives. The idea of computer notebooks has been around for a long time, starting with the early days of Matlab and Mathematica in the mid-to-late-80s. Fast forward 15 years: IPython was just a toddler of a few hundred lines of code, when SageMath became available as a free and open source environment for scientific computing. The past few years haver seen the rise of IPython and its evolution into the Jupyter Project, as well as the emergence of new notebooks, Beaker and Zeppelin. In this article we look at what distinguishes these notebooks and how mature they are. SageMath: The First Open Source Notebook The Sage Notebook was released on 24 February 2005 by William Stein now a professor of mathematics at the University of Washington, as free and open source software (GNU License), with the initial goals of creating an “open source alternative to Magma, Maple, Mathematica, and MATLAB.” Sage is based on Python and Cython and focuses on mathematical worksheets. The Sage Notebook recently moved to the cloud with SageMathCloud, thanks to a collaboration with Google’s cloud services. Sage Notebooks can also be downloaded and run on your local machine. Although not as popular as it ought to be, the Sage Notebook is a free and entirely open source alternative to Mathematica and Matlab that supports Python, LaTeX, Markdown, task lists, R, IPython Notebooks and allows to manage courses, write C programs, make chatrooms, and create Sage worksheets for sophisticated mathematics. The Sage Cloud Notebook can be seen as a hosted version of a platform similar to Matlab or Mathematica that can also embed IPython notebooks. Its math scripting language is as good as its competitors if not better in some cases. And the Sage community is large, deep, and wide with its own stackoverflow type website. The Sage Cloud Notebook works like a charm right out of the box. There’s nothing to install and performances are decent. To test, I simply copied a couple of scikit-learn tutorials into a Sage Notebook, had no issue, got the results as fast as expected, and was able to publish the notebook right away. Sage Cloud is freemium based. With a free plan of 3Gb of storage and 1Gb of memory and 8Gb / 50 Gb for $7/month it offers a pretty competitive deal. It’s a stable environment with few bugs and many online resources and advanced math tutorials. However, with only 20 recent contributors and over 80% of the commits done by William Stein himself, the Sage Notebook could benefit from a larger community of contributors. A bit of visual revamping and better marketing would also go a long way to boost its adoption rates. Jupyter Project Jupyter Notebooks, formerly known as IPython Notebooks, have enjoyed a rather impressive success and steady growth since 2011. In the past year, the number of ipynb files on github has nearly tripled from 80,000 to over 230,000 files. Following the publication of an article on IPython in Nature in Nov 2014, Interactive notebooks: Sharing the code, Rackspace, the company hosting IPython Notebooks, had to ramp up and serve more than 20,000 IPython notebooks. The IPython console was started by Fernando Perez circa 2001. From a first attempt to replicate a Mathematica Notebook with 259 lines code, to the first presentation of the IPython Notebook at EuroSciPy 2011 conference in Paris, IPython had multiple false starts, diverse external influences and time to mature. With the Sage Notebook being a reference all along, Fernando Perez had many collaborations with the Sage team. A detailed history of IPython can be found on Fernando Perez Blog Comparing Google trends for SageMath (red), IPython (Green) and Apache Zeppelin (blue). Note the hockey stick for IPython after the EuroSciPy 2011 conference. In 2015, the IPython Notebook project became the Jupyter project, an ambitious project whose goal is to lay the foundation for a generation of scientific publications focused on reproducibility by making the data and the code accessible and open. The project vision is presented in Project Jupyter: Computational Narratives as the Engine of Collaborative Data Science. See also Brian Granger’s keynote address at ODSC West. As Safia Abdalla writer for Opensource.com puts it “The Jupyter Notebook hints at what the academic journals of tomorrow will look like and paints a promising picture. They will be interactive, visualization-focused, user-friendly, and include code and data as first-class citizens.” The ability to go beyond Python and run several languages in a notebook is also at the center of the Jupyter rebirth. Multilingualism is still limited by notebook. It is not possible to have multiple cells with multiple languages within the same notebook. Furthermore, in order to run notebooks in languages other than Python, you still need to install additional kernels. See this article for a detailed walkthrough of Implementing an R kernel. The Jupyter Project benefits from a large community of contributors, partnerships with many companies (Rackspace, Microsoft, Continuum Analytics, Google, Github, …) and universities (UC Berkeley, George Washington University, NYU, …) Installation is straightforward, support is widely available for all platforms and there is a very active stack overflow stream for all things tagged IPython or Jupyter. There’s a New Notebook in Town In fact not one but two new notebooks have recently blipped on the data scientist radar: the Apache Zeppelin Notebook and the Beaker Notebook. The Zeppelin Notebook The Zeppelin Notebook is supported and incubated by the Apache software foundation with Lee Moon Soo as its lead developer. It is similar in concept to the Jupyter Notebook with several noticeable differences. Java Apache Zeppelin is build on the JVM while Jupyter is built on Python. This can be an obstacle for someone not familiar with the Java ecosystem. (my case in fact). Since the project is in its infancy, it does not yet offer a binary install file. You have to clone the github repository and build it from the source with dependencies such as Open JDK, Maven and Node. The main problem people encounter during the install seem to be with the web front end installation. Trevor Grant has written a very detailed walkthrough to install Zeppelin on Ubuntu 14. See also this tutorial by Bill Chambers and Zeppelin Notebook Tutorial Walkthrough by Tyler Mitchell. Multilingualism Zeppelin offers the possibility to mix languages across cells. Zeppelin currently supports Scala (with Apache Spark), Python (with Apache Spark), SparkSQL, Hive, Markdown and Shell. And you can make your own language interpreter. Spark Apache means Apache Spark. Zeppelin is fully oriented for Spark. It is data exploration and visualization intended for big data and large scale projects. Of course you can use pyspark in a Jupyter Notebook, but Zeppelin is natively Spark. Being part of the Apache ecosystem does not hurt either. Front-End Zeppelin has a more advanced set of front-end features than Jupyter. A lot of efforts have been spent on the Javascript UI which is developed in Angular. It is, for instance, possible to modify the notebook DOM with angular. The content of one cell is modified downstream by another cell. Fun! See a simple demo by Lee Moon Soo of Zeppelin Angular display system. Zeppelin runs a scala interpreter (REPL), which gives you direct access to the DOM. It also has an angular interpreter which allows you to build great UIs and import whole chunks of web technologies. This Zeppelin Notebook offers a good sample of Visualizations In Zeppelin. Although still pretty much in beta, the Zeppelin Notebook enjoys quite a buzz and has a rather large community. The roadmap for version 0.6 includes R support with the implementation of the sparkR interpreter, better Python and angular repl and job management, among other things. The Spark Notebook is an alternative to Apache Zeppelin. It is also based on JVM and offers a mix of languages per notebook. There’s a pretty thorough documentation in the github repo. The Beaker Notebook Scott Draves, the project lead developer, made quite an impression at the recent ODSC Boston Data Festival 2015 conference with his presentation of the Beaker Notebook. Built on top of the IPython kernel and recently open-sourced by Two Sigma, an investment manager firm based in NYC, the Beaker Notebook was designed from the start to be a fully polyglot notebook. It currently supports Python, Python3, R, Julia, JavaScript, SQL, Java, Clojure, HTML5, Node.js, C++, LaTeX, Ruby, Scala, Groovy, Kdb. See for instance this translation notebook example where different types of variables (scalar, dataframes, images, tables, etc.) are set and accessed in Python, R, JS, and Groovy. The Beaker Notebook app runs on Mac, Linux, and Windows, and the install is straightforward. It comes bundled with the usual suspects, markdown, latex and Python. Other languages must be hooked up to be recognized by the notebook which can be challenging and time consuming. Installing a piece of software can always be difficult. This might also be true of Zeppelin or Jupyter Notebooks. What was troubling with the Beaker Notebook was the absence of a vibrant online community. Issues and problems are mostly dealt with through the github repository issues. And with only a mere 90 questions tagged with ‘beaker‘ on Stack overflow and no mailing list for the users or the developers, troubleshooting your install can be a bit problematic. There are not many Beaker Notebooks available yet besides the ones on the beaker publication server. The Beaker Notebook is a great concept that needs a bit of traction and love from the community to take off. It will be interesting to see how more real-world projects and datasets will fare on this platform. Will the language adaptors start hindering the notebook performances for larger datasets? Conclusion You can use IPython with pyspark or use Python in a Zeppelin Notebook. You can mix Py and Rb and R in Beaker or do some complex math in Sage Notebooks and then import a Jupyter notebook. The possibilities are endless and the momentum is here for notebooks to evolve towards language agnosticism, bigger distributed projects, and better visualization flows. Jupyter is an amazing project that feeds and rides the rising wave of data science. And although it’s hugely popular, there is still room for other amazing projects that will in turn inspire Jupyter. As with any aspect of data science and machine learning, it’s an exciting time to be able to witness these fantastic scientific computing projects come alive. You can read more from Alexis Perrier on his blog or follow him on Twitter @alexip Originally published at opendatascience.com.
Jupyter, Zeppelin, Beaker: The Rise of the Notebooks | Open Data Science
1
jupyter-zeppelin-beaker-the-rise-of-the-notebooks-open-data-science-1348f5eb14fe
2018-03-16
2018-03-16 11:46:35
https://medium.com/s/story/jupyter-zeppelin-beaker-the-rise-of-the-notebooks-open-data-science-1348f5eb14fe
false
1,868
null
null
null
null
null
null
null
null
null
Data Science
data-science
Data Science
33,617
Alexis Perrier
Data Science consultant, NLP, Machine Learning, DeepLearning, Signal processing; PhD @TelecomPTech; Washington DC/Paris France.
60a307d5af11
alexisperrier
468
358
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-08-14
2018-08-14 14:52:12
2018-08-14
2018-08-14 15:01:10
1
false
en
2018-08-14
2018-08-14 15:01:10
2
134a52d2c84
1.50566
0
2
0
Overstating the importance of Artificial Intelligence is difficult. When implemented efficiently, AI holds the capacity to boost your…
5
How Artificial Intelligence is Rewriting the Medical Coding Automation Overstating the importance of Artificial Intelligence is difficult. When implemented efficiently, AI holds the capacity to boost your billing business tenfold. In many cases, AI is the thing that is scaling the business rather than the physical workforce. The question on many business minds is how does AI change the way business is done? To help answer this question we analyzed many billing and coding companies. Below is a summarized version of our findings from the research: Coding and billing is a method through which standard codes are established that categorize patient information records and thus dictate the billing towards insurance companies. The aim is to create a standard billing cost that is determined by the code of the patient record. Unfortunately, this process is facing substantial accuracy challenges. This could be attributed to insufficient documentation, inefficient execution of procedures. As stated in tech emergence, according to the Centers for Medicare & Medicaid Services (CMS), errors resulted in $36.21 billion in improper payments in FY2017. (1) The coding industry suffers a huge setback due to the nature of their audits, which take place towards the end of the revenue cycle. Therefore, even if errors are recognized, it is too late to rectify them since the cost of rectification is usually higher than the initial damage. Within the medical coding and billing industry, it was recently reported that billable codes have now crossed a total number of 70,000+ which subsequently increases the need for medical coders at a significant rate. The medical coding job, when done manually, is complicated and requires a higher amount of workforce since there are only so many accounts every individual can handle efficiently. This is part of the reason that the industry has witnessed several instances of inaccuracies, owed to costly mistakes made while trying to keep up with the ever-increasing new codes that are being established. The need of the hour is to create an agile process that allows the medical coding and billing process to flow seamlessly. To read more:http://bit.ly/2MdFhd0
How Artificial Intelligence is Rewriting the Medical Coding Automation
0
how-artificial-intelligence-is-rewriting-the-medical-coding-automation-134a52d2c84
2018-08-14
2018-08-14 15:01:10
https://medium.com/s/story/how-artificial-intelligence-is-rewriting-the-medical-coding-automation-134a52d2c84
false
346
null
null
null
null
null
null
null
null
null
Healthcare
healthcare
Healthcare
59,511
Nancy Khan
null
1673b7d5fe80
nikosplabs
1
1
20,181,104
null
null
null
null
null
null
0
null
0
f95abfe1ced7
2018-02-12
2018-02-12 11:52:22
2018-03-05
2018-03-05 08:15:39
14
false
en
2018-03-09
2018-03-09 08:09:33
6
134a7b87d0c7
9.204717
3
0
0
Do you know the most important ecommerce trends for 2018?Why do you think these trends are winning the hearts of your customers? And what…
5
Ecommerce trends in 2018: 6 growing trends Do you know the most important ecommerce trends for 2018?Why do you think these trends are winning the hearts of your customers? And what do these trends mean for your company? In this blog post, we will hear about the most popular trends in the field of e-commerce from two experts, the first is Frederik Hyldig from s360, who is Head of the bureau’s Paid Search team. The second is Frederik Nielsen from avXperten, who has been working with online sales for more than 15 years. The two gentlemen give us their take on the top trends based on their expertise in the field of e-commerce and based on the the conversations they have had with their own customers. Three reasons why you should spend 10 minutes reading this blog because it will: This blog will inform you of ● The 6 most important ecommerce trends in 2018 ● 6 examples on how others have implemented these trends ● Why these trends are important for you and your customers #1 Strengthen your products’ ranking on Google When you search for “football boots” on Google, two columns pop up. The column on the left shows the general search results for football boots, while the right column shows the product advertisements.We call the column on the right “Google Shopping.” Example of what Google Shopping looks like for football boots Google Shopping’s advertisements appeal to customers who need a quick overview of a product — here they are able to compare products and prices from different companies with a simple Google search. Google Shopping is for those who want to show products in a clear and inviting manner. Here, your customers can see: a picture of the product from your webshop, the product’s name, the price of the product and the name of your webshop. Why Google Shopping is important Advertisements on Google Shopping raise awareness via the search engine, and these ads also get more attention from the customer than ordinary search results on Google. The illustration shows that Google Shopping has overtaken the text adverts “On the basis of data from our e-commerce customers, we can see that Google Shopping has recently overtaken text adverts and, on average, now account for more than half of the paid traffic from Google to Danish webshops. Just 4 years ago, Shopping was almost non-existent in Denmark. Therefore, if you operate a webshop, Shopping is crucial if you are to get the maximum return from your AdWords advertising.” Frederik Hyldig, Head of Paid Search, s360 #2 Engage your customers “What will the weather be like today?” This is an example of what someone may ask through a voice search. A voice search is when customers use their voices to search for information or to make a purchase. The video here shows how little time it takes for a man to purchase a bar of soap via voice search. Voice search appeals to busy people — or those who are generally on the move a lot. It could also be useful for people who like to multitask, so they can search for something while they are doing something else. Companies that focus on Voice search try to engage their customers by helping them in their busy, day-to-day lives. Exemplary instances of companies that have voice searches are Amazon, Dominos, Whirlpool and Ocado. Alexa, developed by Amazon, could be your new digital and personal assistant. Alexa has more than 1,500 different skills with which you can, for example, get Alexa to set an alarm, send text messages, order food to be delivered, add milk to your shopping list, find the nearest Greek restaurant, play Happier by Ed Sheeran or get it to read a book aloud for you (to name a few…). Domino’s has also made it possible for their customers to order pizza via voice search. Whirlpool lets you start your washing machine, pause it, or adjust the temperature of your oven with a simple voice command. With the Ocado app, you can add “lemons” or other groceries to your shopping list by just saying the missing items out loud. How to ask Oacdo to add lemons to your trolley Why voice searching is important Data from Comscore shows that 50% of all searches in 2020 will come via voice searches. Some people also believe that voice searches are more relevant and purchase-inducing than a normal written inquiry. Adding a voice search component means that your customers will come closer to going through with their purchase. With the traditional text search, there are endless searches on Google that do not lead to any noticeable results for you as a webshop owner. Despite the fact that Voice searches have not yet gained a foothold in Denmark, it is still important to be aware of this trend. “Even though voice searches have not made a breakthrough in Denmark yet, it is already extremely widespread in the USA, and the day products like Amazon Echo and Google Home start to understand Danish, I am certain that it will also be huge in Denmark. When it happens, you, as a webshop owner, must ensure that you also have a presence in this new type of search result.” Frederik Hyldig, Head of Paid Search, s360 #3 Be where your customers are Smartphones have taken control. This does not mean that fewer people are using their computer, it just means that there are more who have started using their phones when they carry out a search online. Therefore, it is crucial that your website is mobile-friendly. When your website is mobile-friendly, it means that your customers can access your website on any size of device without having to zoom in to find the shopping basket or to read the product description. Previously, mobile-friendly websites were created with a responsive design, but a better solution to creating a mobile-friendly website is with Mobile First. Mobile First enables you to design your website on the telephone rather than the computer. Why Mobile First is important Mobile First must be taken seriously by all companies with a website and a webshop who wish to be accessible to their customers on the first part of their shopping journey, where the customer finds inspiration. If you are not in the running when your customers take the first step on their shopping journey, you can’t expect them to choose you when they have selected a webshop they want to buy from. “Our data shows that almost half of the traffic today comes from mobile devices, and this shift means that you have to offer users a positive experience on their mobile phone if you don’t want to see your conversion rate fall.” Frederik Hyldig, Head of Paid Search, s360 The conversion rate will decrease when the traffic moves from the computer to the telephone. Therefore, you should break down any barriers that may be present on the telephone. This can be achieved through more modern payment solutions in your checkout flow, such as MobilePay, ViaBill, and Apple Pay. If you would like to find out more about why Mobile First is important and how to get started with it, then read this earlier blog post about mobile first. #4 Be accessible “here and now” Websites often have a live chat function, with the main purpose of giving customers an answer “here and now.” The chat box normally pops up as soon as the customer accesses the website. On some websites, the chat function is set so it appears when the customer has been active on the website for a certain amount of time. The chat function is used by customers who want to find out the price of an item or are looking for a specific product. It is easier for the customer to send a quick message in the chat box than to call a representative or send an e-mail. This is an appealing function to customers because they receive an answer to their question instantaneously, when they are in the middle of a purchase on the website. 3 reasons why customers prefer chat boxes If you, as companies, want to be accessible when your customers have questions, even at 10 o’ clock on a Friday evening, then the chat box is an option. Chat boxes are for those in particular who have many products or technical products on their website. Why a chat box is important When your customers need an answer in order to proceed on their shopping journey, a chat box can give them a swift answer. If they don’t get an answer or a swift answer when they are on your website, then you run the risk of losing them because they specifically need an answer in order to take the final step on their shopping journey. “We are seeing a higher conversion rate when we have the chat function activated. And we are getting some pre-sale inquiries that we otherwise wouldn’t have gotten (we measure this by the fact that we get more inquiries when we have chat activated — if we measure the number of chats + telephone calls versus the number of telephone calls when chat is deactivated — then we get more inquiries. In other words, there are inquiries we don’t get when chat is deactivated because they don’t come via telephone calls.” Frederik Nielsen, CEO, avXperten #5 Customise your products Machine learning is used by companies who have the confidence to be brave enough to give their systems access to their customers’ data, so that the systems themselves can use this data and learn from it. This means that the recommendations that appear on the screen are not programmed by humans, but by a black box with an algorithm. Thus, machine learning gets the computer to do something without being explicitly programmed to do it. “In the last decade, machine learning has given us self-driving cars, speech recognition, effective web searching and much more. Machine learning is so widespread today that you probably use it several times every day without knowing it.” Frederik Nielsen, CEO, avXperten Examples of recognisable companies that have implemented machine learning are Amazon, Netflix and Spotify. Amazon has, for example, made it possible for their customers to be presented with new recommendations. These recommendations are found on the basis of the customer’s previous purchases or the books the customer currently has on their bookshelf, or the items in one’s shopping cart. On Spotify, you have probable become acquainted with “Discover Weekly”. This is also an example of machine learning. Here, Spotify’s algorithms find a personal playlist composed of different songs you haven’t heard before. How Discover Weekly looks like on Spotify This is why machine learning is important More than 80% of all the series and films shown on Netflix have been put together via Netflix’s own recommendation system. What does this tell us? It tells us that machine learning is essential if you want to find products that match your customers’ previous purchases. The more the products match the customer’s preferences, the more inclined the customer will be to buy them. Machine learning can be used to give your customers a tailor-made experience when they shop in your webshop or website. #6 Beware of Amazon Amazon is spearheading almost half of all online sales in the U.S., which is something all companies should be aware of. There are also several indications that Amazon is making advances towards the Nordic countries. Read more here. Stats about Amazons online growth How to handle Amazon It may be the large generalist webshops that will feel the Amazon effect the hardest, as they are not able to provide detailed advice on their products but are more a sales platform for an extremely broad and deep range of goods. Therefore, Danish companies should start doing what Amazon will never be able to — that is, give customers a knowledgeable customer service with regard to the products. It is the customer service, and the specialised advice about the company’s products that separates Amazon from the individual niche stores. You have now been given an introduction to 6 of the most popular trends in 2018 according to Frederik Hyldig’s and Frederik Nielsen’s assessment. But, we are also keen to know which trends in the field of e-commerce you find current for this year, and why. If you would like to give your customers a greater incentive to purchase on the telephone, get in touch with ViaBill at this e-mail address: kbc@viabill.com
Ecommerce trends in 2018: 6 growing trends
18
ecommerce-trends-in-2018-6-growing-trends-134a7b87d0c7
2018-04-10
2018-04-10 22:07:08
https://medium.com/s/story/ecommerce-trends-in-2018-6-growing-trends-134a7b87d0c7
false
2,055
We want to help you to make online shopping easy and flexible for your customers. And we want to give you specific advice on how to choose the most important things to focus on at your website.”
null
null
null
ViaBill.com
null
viabill-dk
ONLINE BUSINESS,ONLINE SHOPPING,PAYMENTS,CONVERSION RATE
null
Voice Search
voice-search
Voice Search
345
ViaBill
We want to make online shopping as easy and transparent as possible for you and your customers.
ead12abdcbc9
viabill
8
5
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-05-22
2018-05-22 14:07:38
2018-05-22
2018-05-22 04:53:25
4
false
en
2018-05-22
2018-05-22 14:07:39
12
134aa81c37ac
4.75283
0
0
0
null
5
Cisco and BroadSoft: Life After the Acquisition When Cisco, one of the world’s leading enterprise communication companies, announced their decision to purchase cloud specialists BroadSoft, the news sent shockwaves around the UC marketplace. In February of this year, the deal came to a successful close, bringing BroadSoft’s exceptional cloud strategy and collaboration services into Cisco’s leading meetings, messaging, calling, and customer care portfolio. At UC EXPO this week, we had anopportunity to catch up with Craig Decker, the Managing Director for EMEA at BroadSoft (now part of Cisco). With more than 11 years working for BroadSoft, Craig has been in a fantastic position to witness the merger of the two companies. We sat down with a quick discussion with him about how the merger has affected the two businesses, and what the future looks like for Cisco and BroadSoft. What’s Been Happening Since the Deal? It’s been a few months since the ink dried on the acquisition agreement between Cisco and BroadSoft, making Cisco one of the biggest, and most compelling UC solutions in the current marketplace. I asked Craig what’s been happening since the deal took place, and he told me that the transformation has been significant. “Ever since the early stages of the acquisition, it was clear that Cisco wanted to hold onto the BroadSoft DNA and the way we approached managing service providers. Our integration into the Cisco Collaboration Group has been about delivering collaboration and cloud calling to customers and offering sales and marketing support to service providers using the Cisco portfolio.” As Decker told me, BroadSoft has always been 100% service provider focused. “We were never really a brand that enterprise companies knew about. I think one of the biggest things that have changed since the acquisition, is that we now have countless enterprise providers coming to work with us. Enterprise service providers are seeing a huge opportunity in the merger because even though Cisco has always been a market leader in terms of enterprise communications, they never really had that cloud calling aspect. With BroadSoft, they can really offer that to their customers.” Decker noted that the best thing about the acquisition for him is how welcoming Cisco have been. “As an integration company, Cisco is world class. My team has felt so welcome, and the process has been incredibly smooth. I think there’s always a risk of people losing focus during change, but I’ve seen the reverse. Our people can see the opportunity, and they’re so motivated.” Where Do You Think the Market Transformation in UC is Heading? Now that Cisco and BroadSoft have combined their specialties into one fantastic portfolio, Craig believes that they’re going to see a huge expansion in market reach, particularly when it comes to mid-market and enterprise customers. “I think there’s a lot of opportunity, but we’ll need to educate all the different channels about what’s possible first. It’s all about creating the right transitional experience.” Of course, it’s not just Cisco, BroadSoft, and their customers that are going through a transformation right now. The whole UC market is in a state of flux, with plenty of consolidation and growth throughout. I asked Craig where he thinks the market is heading. “I think we’re going to see continued service provider mergers and continued new developments by vendors like Cisco as everyone starts to deliver more complete portfolios. There’s a lot of vendors out there today that are pure cloud — and while that might work for SMBs that can make the jump to cloud pretty quickly, there needs to be more hybrid opportunities for mid-market and large enterprises too.” Decker also commented that as the move to cloud accelerates, he thinks that we’ll start to see more specialised and niche solutions come up across various verticals. “I think the market segment will continue to evolve as people solve the different challenges that various verticals face.” Will the Cisco and BroadSoft Collaboration Stories Be Merging Together? Craig Decker, Managing Director EMEA, BroadSoft One discussion that took place at the UC EXPO this year, was all about how Cisco’s collaboration story is evolving. Of course, before the acquisition, BroadSoft had their own collaboration story too. I asked Craig whether he thought the two would be coming together or not. “If you look at what BroadSoft was, and what it is, it was always the white-label solution that service providers would take to suit their needs. They would go to market with their own branded service, provided by BroadSoft. In the SMB market — this is sure to continue. Cisco has been very clear that the APIs and openness we had in allowing carrier branded solutions will continue in this SMB section. What we’re doing in the collaboration group is taking the best of that, and our calling platform, along with Cisco’s branded components, and making it easier for service providers to consume in the cloud.” Cisco and BroadSoft are both considering the nature of the brands they work with, and their route to market, with a complete portfolio that spans everything from the SMB to the large enterprise. “We expect to continue offering service providers the best of both worlds from BroadSoft and Cisco. For small companies, that might be mostly BroadSoft services, for mid-market, it might be a combination of Cisco and BroadSoft, and for large enterprise, it could be mostly Cisco. How Does the Future Look for Cisco and BroadSoft? Cisco VR With so many discussions about disruptive technology like AI and IoT emerging at UC EXPO, I was keen to hear about what the future might look like for BroadSoft and Cisco. Craig told me: “One of the things we’re looking at in AI and contextual search and so on, is that we’ve got so much information now that sits in the cloud. The cloud gives us a unique opportunity to run nurture programs and analytics that allow us to create a smarter environment for customers, and I think you’ll see a lot of focus in the coming years on taking AI from command and control, to being a part of improving usability.” Decker also noted that he believes we’ll start to see more contextual capabilities brought into systems to make managing communication and collaboration easier. “It’s about making sure that the technology helps you, rather than hindering you, and I think Cisco is already ahead of the curve. It’s bringing the pieces of the puzzle together and delivering amazing solutions in a meaningful way.”
Cisco and BroadSoft: Life After the Acquisition
0
cisco-and-broadsoft-life-after-the-acquisition-134aa81c37ac
2018-05-22
2018-05-22 14:07:41
https://medium.com/s/story/cisco-and-broadsoft-life-after-the-acquisition-134aa81c37ac
false
1,074
null
null
null
null
null
null
null
null
null
Analytics
analytics
Analytics
15,193
UC Today
Unified Communications Stories
bd51979d153c
uctoday
13
74
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-04-21
2018-04-21 07:22:29
2018-04-22
2018-04-22 18:53:20
2
false
en
2018-04-22
2018-04-22 18:53:20
5
134b5b330e95
2.522956
23
1
0
Mask RCNN has been the new state of art in terms of instance segmentation. There are rigorous papers, easy to understand tutorials with…
5
Simple Understanding of Mask RCNN Mask RCNN has been the new state of art in terms of instance segmentation. There are rigorous papers, easy to understand tutorials with good quality open source codes around for your reference. Here I want to share some simple understanding of it to give you a first look. Source: Mask RCNN paper Mask RCNN is a deep neural network aimed to solve instance segmentation problem in machine learning or computer vision. In other words, it can separate different objects in a image or a video. You give it a image, it gives you the object bounding boxes, classes and masks. There are two stages of Mask RCNN. First, it generates proposals about the regions where there might be an object based on the input image. Second, it predicts the class of the object, refines the bounding box and generates a mask in pixel level of the object based on the first stage proposal. Both stages are connected to the backbone structure. What is backbone? Backbone is a FPN style deep neural network. It consists of a bottom-up pathway , a top-bottom pathway and lateral connections. Bottom-up pathway can be any ConvNet, usually ResNet or VGG, which extracts features from raw images. Top-bottom pathway generates feature pyramid map which is similar in size to bottom-up pathway. Lateral connections are convolution and adding operations between two corresponding levels of the two pathways. FPN outperforms other single ConvNets mainly for the reason that it maintains strong semantically features at various resolution scales. Now let’s look at the first stage. A light weight neural network called RPN scans all FPN top-bottom pathway( hereinafter referred to feature map) and proposes regions which may contain objects. That’s all it is. While scaning feature map is an efficient way, we need a method to bind features to its raw image location. Here come the anchors. Anchors are a set of boxes with predefined locations and scales relative to images. Ground-truth classes( only object or background binary classified at this stage) and bounding boxes are assigned to individual anchors according to some IoU value. As anchors with different scales bind to different levels of feature map, RPN uses these anchors to figure out where of the feature map ‘should’ get an object and what size of its bounding box is. Here we may agree that convolving, downsampling and upsampling would keep features staying the same relative locations as the objects in original image, and wouldn’t mess them around. At the second stage, another neural network takes proposed regions by the first stage and assign them to several specific areas of a feature map level, scans these areas, and generates objects classes(multi-categorical classified), bounding boxes and masks. The procedure looks similar to RPN. Differences are that without the help of anchors, stage-two used a trick called ROIAlign to locate the relevant areas of feature map, and there is a branch generating masks for each objects in pixel level. Work completed. Illustration of Mask RCNN structure The most inspiring things I found about Mask RCNN is that we could actually force different layers in neural network to learn features with different scales, just like the anchors and ROIAlign, instead of treating layers as black box. If you are interested in more implementation details of Mask RCNN, please read the links given in this article. Leave comments below if you have any questions.
Simple Understanding of Mask RCNN
57
simple-understanding-of-mask-rcnn-134b5b330e95
2018-06-20
2018-06-20 19:25:23
https://medium.com/s/story/simple-understanding-of-mask-rcnn-134b5b330e95
false
567
null
null
null
null
null
null
null
null
null
Object Detection
object-detection
Object Detection
320
Xiang Zhang
null
ac88a473230b
alittlepain833
17
5
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-10-24
2017-10-24 17:47:50
2017-10-24
2017-10-24 20:41:31
1
false
en
2017-10-24
2017-10-24 20:41:31
16
134bda7876a0
3.713208
4
0
0
I had the privilege of attending the Rock Health Summit last week in San Francisco, where healthcare startups, payers, providers, and…
5
Three Takeaways from the Rock Health Summit 2017 I had the privilege of attending the Rock Health Summit last week in San Francisco, where healthcare startups, payers, providers, and government officials converged and talked about all things digital health. A few key takeaways from this year’s event: 1. No more tech platform point solutions, please This sentiment was pervasive: the digital health landscape today is full of tech platform point solutions that were built without fully understanding the problems their clients are trying to solve. Also, many panelists pointed out that successful digital health companies often go beyond providing just a “tech platform”, but also offer a full suite of services supplementing the “tech”, thinking in terms of the complex matrix of healthcare stakeholders, their existing workflows and processes, and the combination legacy and new technologies. For example, Flatiron Health offers a full oncology EMR in addition to their analytics product for their clients, while Omada Health employs and trains health coaches to interact directly with patients on top of their platform. “AirBnb doesn’t call themselves a ‘tech’ company, they just use technology to get stuff done,” said Andy Slavitt, former CMS Administrator under the Obama administration, “we need to stop with this fascination over technology in healthcare and fully understand the problems we are trying to solve.” 2. How to parse out the game-changing companies through the noise Many believe that the B2C business model within healthcare is a thing of the past, though, as Yumin Choi of Bain Capital Ventures pointed out, B2B2C models can be successful (e.g., WebMD with alternative revenue sources). So where do all these B2B companies turn to? “Follow the money” — payers and self-insured employers (who are no longer a secret distribution channel). Having a tangible, demonstrated ROI is imperative, however. In an environment where payers, self-insured employers, and VCs are constantly inundated with solutions claiming to be the silver bullet, CFOs and Investment Managers alike need to justify the spend or investment on a healthcare solution. Savvy investors and buyers will also go beyond the “shiny growth metrics” and deep-dive on the results of sophisticated clinical trials. For example, Virta Health was able to turn heads with a peer-reviewed clinical trial that showed the potential of reversing Type 2 diabetes. If organizations can demonstrate both economic validation (does it sell?) along with clinical validation (does it work?), they could be a major game changer. 3. Yes, AI is over-hyped, but companies properly gearing up will win “It’s hard to differentiate between the companies that are truly innovative with AI vs. those that have an excel spreadsheet,” says Jessica Mega, Chief Medical Officer at Verily, perfectly summing up the current over-hype fueled by the sheer number of companies touting AI and all things AI related. But even with the over-hype, there is general consensus that AI will be transformative in many aspects in healthcare. So what are the leaders in AI doing today? It turns out they are investing in the broader infrastructure and architecture to first tackle the problems that come with large, fragmented sources of data, which at its core, are actually governance and organizational problems. These problems manifest themselves such as: How to aggregate data from multiple legacy/new sources? How to clean data efficiently? How to correctly label the data for algorithms to ingest, and what if there are disagreements relative to the necessary labeling? Andrew Ng shares his thoughts on what makes successful health AI companies Echoing the need to focus on the core infrastructure is AI luminary / Co-Founder of Coursera Andrew Ng. Andrew argued that companies that strategically gather data, centralize their data in a data warehouse, and recruit the best AI talent (see his HBR article: Hiring your first Chief AI Officer), are going to have a leg up in this AI race. From an organizational point of view, Andrew argues that there needs to be a centralized AI team/unit that is matrixed into the different business units within a larger organization. He draws the comparison to successful companies that built mobile functions within different business units during the mobile revolution a decade ago. From a regulatory perspective, successful companies are working closely with and/or keeping an eye on the comings and goings of regulatory agencies (e.g., FDA) to prove out that AI is safe and effective. Bakul Patel, the Associate Director for Digital Health of FDA, urged startups to keep engaged with the FDA through its website and webinars to better understand the FDA’s areas of interest as well as priorities. Lastly, panelists agree: at the end of the day, AI is just a tool, and without clearly defining the problem you’re trying to solve, these fancy AI algorithms are meaningless. Shameless Plug Here at Spreemo Health, we are focused on addressing one of the most fundamental problems in healthcare: in a world where care quality is highly variable, how do you identify the right doctor to give you the greatest chance of getting better? We’re solving this problem starting in radiology, utilizing our proprietary, labeled datasets of advanced imaging, medical reports, and partnering with our payer-clients to identify and measure the impact of diagnostic errors. Utilizing these insights, we have created a nationwide diagnostic network of centers of excellence on a zip-code to zip-code level, ensuring that providers are adhering to strict guidelines and protocols for Spreemo Health patients, whether in New York, NY or Newport Beach, CA. I’m always looking to connect with others within the industry working on tackling interesting problems in healthcare. Please feel free to reach out to me via LinkedIn!
Three Takeaways from the Rock Health Summit 2017
10
three-takeaways-from-the-rock-health-summit-2017-134bda7876a0
2018-06-08
2018-06-08 16:47:06
https://medium.com/s/story/three-takeaways-from-the-rock-health-summit-2017-134bda7876a0
false
931
null
null
null
null
null
null
null
null
null
Healthcare
healthcare
Healthcare
59,511
Julian Yao
I like all things #healthcare #data #infographics #airplanes | Product @SpreemoHealth | Previously Consultant @Accenture
a56ac9cc9b52
julianyao
74
102
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-04-02
2018-04-02 21:15:30
2018-04-02
2018-04-02 21:24:49
1
true
en
2018-04-02
2018-04-02 21:24:49
2
134d729dcf2a
2.120755
0
0
0
In the future, you won’t have to take care of your old parents- something else will do that for you.
5
Something’s wrong with Technology Photo by Alex Knight on Unsplash In the future, you won’t have to take care of your old parents- something else will do that for you. Although I won’t throw my gears away nor will I encourage anyone to do that, here’s my concern as time wave bye- what’ll happen to human existence? The world is gradually losing a feel of human existence. We’re gradually losing everything up to technology. We’ve forgotten the initial aim of technology “to aid human activities”. But today, humans are gradually losing: · The joy of togetherness · The joy in a fist fight to prove our strength · The dignity of manual labor Now, lethal weapons are used for mass destruction to prove strength. We’re now satisfied with video calls and chats instead of hurrying home from work to see our loved ones Job scarcity everywhere because of technological involvement- yet we call it advancement. Have you ever wondered who’ll take care of your old parents as they grow older- You or machines? Many people have given up to robots becoming our caretakers. In places like Japan, humans have concluded there won’t be enough young ones to take care of their older parents and now inventing machines “robots” that can sing or speak when touched. Technology has given us a new sense of care. In the past, we used to believe that care means CARING about someone but it’s different today. We’ve changed the context of care to taking care of someone which in reality, a robot can do. We’ve lost the feel of the care a human or robot can give. We don’t even believe there’s a difference anymore. Now, humans have stupidly believed that machines could actually measure the length of their love for a partner or a partner’s love for them. Isn’t that insanity? Technology isn’t inside our hearts, blood or bone marrow. What happened to those times we show strength with fist fights? We now use guns, bombs, missiles and other lethal mass destruction weapons. What happened to those times we visit our kids in school or go through their evaluated their performance by observation and physical examination? We now use gears to monitor our kids. What happened to those times we travelled some miles to see our old parents and show them some care for all the love they showered on us when we were still tender and weak? We leave them in the care of machines called “care giving robots”. What happened to those times partners were satisfied with sexual feelings from their lovers? We now take pleasure in sexual affairs with robots or toys. Now I’m scared… Very soon there’ll be a war between the animate and inanimate. There’ll be a fierce war between Humans and robots. As robots become more humanlike, they’ll definitely hurt us emotionally. By the time Robots ask us to care for them too then, we’ll see the blurring of the line between the animate and inanimate. We still have a little time now. We can take our place again from the threatening technology.
Something’s wrong with Technology
0
somethings-wrong-with-technology-134d729dcf2a
2018-04-03
2018-04-03 13:05:49
https://medium.com/s/story/somethings-wrong-with-technology-134d729dcf2a
false
509
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Lilywriter
null
139ef76e60c
LIlyMichael
45
95
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-08-21
2018-08-21 05:14:49
2018-08-21
2018-08-21 05:33:09
0
false
en
2018-08-21
2018-08-21 05:52:59
7
134eb7c73143
0.249057
0
0
0
Presenting to you the exciting Rap MV (in Mandarin Chinese) made by our very own energetic, creative and professional BIBOX team:
5
First ever Crypto exchange Rap MV — Bibox is Here! Presenting to you the exciting Rap MV (in Mandarin Chinese) made by our very own energetic, creative and professional BIBOX team: Look out for our Founder and China Country Manager Jeffery Lei who also rapped in the video! Further References for Bibox Exchange: Website: https://www.bibox.com Telegram: https://t.me/biboxglobal, https://t.me/Bibox_News_EN, https://t.me/Bibox_News_CN Facebook: https://www.facebook.com/Bibox-612397599100700/ Twitter: https://twitter.com/Bibox365 Medium: https://medium.com/@biboxexchange
First ever Crypto exchange Rap MV — Bibox is Here!
0
first-ever-crypto-exchange-rap-mv-bibox-is-coming-134eb7c73143
2018-08-21
2018-08-21 05:52:59
https://medium.com/s/story/first-ever-crypto-exchange-rap-mv-bibox-is-coming-134eb7c73143
false
66
null
null
null
null
null
null
null
null
null
Blockchain
blockchain
Blockchain
265,164
Bibox Exchange
Bibox is an AI-enhanced encrypted digital asset exchange and is one of the largest global exchanges.
b6ebb5fff59c
biboxexchange
63
1
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-12-01
2017-12-01 11:24:38
2017-11-30
2017-11-30 09:41:20
1
false
en
2017-12-01
2017-12-01 11:39:21
2
134f074b15fd
6.056604
2
0
0
Many people see machine learning as a path to artificial intelligence, but for a statistician or a businessman, it can also be a powerful…
1
Introduction to Machine Learning Many people see machine learning as a path to artificial intelligence, but for a statistician or a businessman, it can also be a powerful tool allowing the achievement of unprecedented predictive results. Why Machine Learning is so Important Before we start learning, we would like to spend a few minutes emphasizing WHY machine learning is so important. Everyone knows about artificial intelligence or AI in short. Usually, when we hear AI, we imagine robots going around, performing the same tasks as humans. But we have to understand that, while some tasks are easy, others are harder, and we are a long way from having a human-like robot. Machine learning, however, is very real and is already here. It can be considered a part of AI, as most of what we imagine when we think about an AI is machine learning based. In the past, we believed these robots of the future would need to learn everything from us. But the human brain is sophisticated, and not all actions and activities it coordinates can be easily described. Arthur Samuel, in 1959, came up with the brilliant idea that we don’t need to teach computers, but we should rather make them learn on their own. Samuel also coined the term “machine learning”, and since then, when we talk about a machine learning process, we refer to the ability of computers to learn autonomously. Examples of Machine Learning While preparing the contents of this post, I wrote down examples with no further explanation, presuming everyone is familiar with them. And then I thought: do people know these are examples of machine learning? Let’s consider a few. Natural language processing, such as translation. If you thought Google Translate is a really good dictionary, think again. Oxford and Cambridge are dictionaries that are constantly improved. Google Translate is essentially a set of machine learning algorithms. Google doesn’t need to update Google Translate; it is updated automatically based on the usage of different words. Oh, wow. What else? While still on the topic, Siri, Alexa, Cortana, and recently Google’s Assistant are all instances of speech recognition and synthesis. There are technologies that allow these assistants to recognize or pronounce words they have never heard before. It is incredible what they can do now, but they’ll be much more impressive in the near future! What else? SPAM filtering. Unimpressive, but it is noteworthy that SPAM no longer follows a set of rules. It has learned on its own what is SPAM and what isn’t. Recommendation systems. Netflix, Amazon, Facebook. Everything that is recommended to you depends on your search activity, likes, previous behavior, and so on. It is impossible for a person to come up with a recommendation that will suit you as well as these websites do. Most important, they do that across platforms, across devices, and across apps. While some people consider it intrusive, usually, that data is not processed by humans. Often, it is so complicated that humans cannot grasp it. Machines, however, match sellers with buyers, movies with prospective viewers, photos with people who want to see them. This has improved our lives significantly. If somebody annoys you, you won’t see that person popping up in your Facebook feed. Boring movies rarely make their way into your Netflix account. Amazon is offering you products before you know you need them. Speaking of which, Amazon has such amazing machine learning algorithms in place they can predict with high certainty what you’ll buy and when you’ll buy it. So, what do they do with that information? They ship the product to the nearest warehouse, so you can order it and receive it in the same day. Incredible! Machine Learning for Finance Next on our list is financial trading. Trading involves random behavior, ever-changing data, all types of factors from political to judicial that are far away from traditional finance. While financiers cannot predict much of that behavior, machine learning algorithms take care of that and respond to changes in the market faster than a human can ever imagine. These are all business implementations, but there are even more. You can predict if an employee will stay with your company or leave. You can decide if a customer is worth your time, if they’ll likely buy from a competitor or not buy at all. You can optimize processes, predict sales, discover hidden opportunities. Machine learning opens a whole new world of opportunities, which is a dream come true for the people working in a company’s strategy department. Anyhow, these are uses already here. Then we have the next level, like autonomous vehicles. Machine Learning Algorithms Self-driving cars were science fiction until recent years. Well, not anymore. Millions, if not billions, of miles have been driven by autonomous vehicles. How did that happen? Not by a set of rules. It was rather a set of machine learning algorithms that made cars learn how to drive extremely safely and efficiently. We can go on for hours, but I believe you got the gist of: “Why machine learning”. So, for you, it is not a question of why, but how. That’s what our Machine Learning course is tackling. How to create machine learning algorithms! That’s really exciting! The focus of this course is deep learning and deep neural networks in particular. It is of upmost importance to us to provide you with an in-depth preparation. We don’t want to scratch the surface; we want to enable you to see the full picture of how things are done. We will cover very hot and trendy topics already shaping the future, and we are super excited about that! Creating a Machine Learning Algorithm Creating a machine learning algorithm ultimately means building a model that outputs correct information, given that we’ve provided input data. For now, think of this model as a black box. We feed input, and it delivers an output. For instance, we may want to create a model that predicts the weather tomorrow, given meteorological information for the past few days. The input we’ll feed to the model could be metrics, such as temperature, humidity, and precipitation. The output we will obtain would be the weather forecast for tomorrow. Now, before we get comfortable and confident about the model’s output, we must train the model. Training is a central concept in machine learning, as this is the process through which the model learns how to make sense of the input data. Once we have trained our model, we can simply feed it with data and obtain an output. Training a Machine Learning Algorithm The basic logic behind training an algorithm involves four ingredients: data model objective function and an optimization algorithm Let’s explore each of them. First, we must prepare a certain amount of data to train with. Usually, this is historical data, which is readily available. Second, we need a model. The simplest model we can train is a linear model. In the weather forecast example, that would mean to find some coefficients, multiply each variable with them, and sum everything to get the output. As we will see later, though, the linear model is just the tip of the iceberg. Stepping on the linear model, deep machine learning lets us create complicated non-linear models. They usually fit the data much better than a simple linear relationship. The third ingredient is the objective function. So far, we took data, fed it to the model, and obtained an output. Of course, we want this output to be as close to reality as possible. That’s where the objective function comes in. It estimates how correct the model’s outputs are, on average. The entire machine learning framework boils down to optimizing this function. For example, if our function is measuring the prediction error of the model, we would want to minimize this error or, in other words, minimize the objective function. Our final ingredient is the optimization algorithm. It consists of the mechanics through which we vary the parameters of the model to optimize the objective function. For instance, if our weather forecast model is: Weather tomorrow equals: W1 times temperature, plus W2 times humidity, the optimization algorithm may go through values like: W1 and W2 are the parameters that will change. For each set of parameters, we would calculate the objective function. Then, we would choose the model with the highest predictive power. How do we know which one is the best? Well, it would be the one with an optimal objective function, wouldn’t it? Alright. Great! Did you notice we said four ingredients, instead of saying four steps? This is intentional, as the machine learning process is iterative. We feed data into the model and compare the accuracy through the objective function. Then we vary the model’s parameters and repeat the operation. When we reach a point after which we can no longer optimize, or we don’t need to, we would stop, since we would have found a good enough solution to our problem. In the first section of our Machine Learning course , we look at each of the basic building blocks in more detail. At the end, we conclude by creating our first machine learning algorithm! Sounds exciting? Well, it definitely is! Originally published at 365datascience.com on November 30, 2017.
Introduction to Machine Learning
51
introduction-to-machine-learning-134f074b15fd
2018-05-11
2018-05-11 03:39:05
https://medium.com/s/story/introduction-to-machine-learning-134f074b15fd
false
1,552
null
null
null
null
null
null
null
null
null
Machine Learning
machine-learning
Machine Learning
51,320
365 Data Science
https://365datascience.com is an educational career website, designed for aspiring marketing analysts, BI analysts, data analysts and data scientists.
e960cfb4e73c
team_31807
43
1
20,181,104
null
null
null
null
null
null
0
null
0
661161fab0d0
2018-08-27
2018-08-27 17:58:13
2018-08-27
2018-08-27 18:40:08
1
false
en
2018-08-28
2018-08-28 01:03:29
0
134f9fa24487
2.430189
8
0
0
One of the famous scientist of our times Stephen Hawking said about Artificial Intelligence — “AI is likely to be either the best or worst…
5
Why you should discuss Artificial Intelligence with your child ? One of the famous scientist of our times Stephen Hawking said about Artificial Intelligence — “AI is likely to be either the best or worst thing to happen to Humanity”. We had Agricultural revolution five thousand years ago , Industrial revolution around 300 years ago , then Internet revolution and now we are heading towards another revolution -Super Intelligent Machines Revolution. Do you think this will be the last revolution by humans ? I will keep this question for you :).These days AI is the buzzword everywhere. We all know Artificial Intelligence will have a big impact on the Human society in coming years. AI has the potential to change the way we use healthcare , financial services ,education ,entertainment, governance and the list goes on. AI will solve some real problems which we have not been able to solve in the past. My aim is not to discuss the benefits of AI in this article ,We all know how it will benefit the world if used wisely. One thing we have to accept that our children will grow up in the AI enabled world.Most of us think about the skills like critical thinking , coding, learning algorithms ,analytical ability which will be needed by our children in AI enabled world. Fair enough ,but I want to talk about the other aspect of AI which we should think when our children are growing up in AI enabled world. Do you have answers to all your child’s questions ? No you don’t but Alexa or Siri can help them in their all questions or doubts ? Siri, Alexa or your child’s AI enabled toy will know more about your child likes and dislikes than you . It is possibility that your child can form emotional bond with these bots.These toys can be very influential because the children will be conversing with them all the time . Children will trust these machines more than you because these machines will listen to them every time. What if your child will doubt his own intelligence because he will think robots are more intelligent and smarter than him? Your child won’t need any friends because he will already have one who will listen to him or her every time. With the advancements of these AI bots , Will it be possible for your child to differentiate between the machine and human feelings ? Don’t you think this will eventually make children to live in their own world ?They will not get time to learn human values and relationships. I think we must prepare our children for the new challenges and opportunities posed by AI. We should make our children understand what it feels to be human. Our children should not become slaves to these machines . We should discuss about the dangers and benefits of AI with them. This is not to scare them but to make them realize that humans have created these machines not the reverse. Our children needs to use the technology wisely for their better future. Our children will inherit the world from us so its our duty to make them aware of the things which will impact them in future. They should develop the machines which should be used for the betterment of human society. They should understand that machines cannot teach us human values and ethics. The curiosity and knowledge of our children should work for the development of responsible AI which should serve the humanity in a good way.
Why you should discuss Artificial Intelligence with your child ?
245
why-you-should-discuss-artificial-intelligence-with-your-child-134f9fa24487
2018-08-28
2018-08-28 01:03:30
https://medium.com/s/story/why-you-should-discuss-artificial-intelligence-with-your-child-134f9fa24487
false
591
where the future is written
null
null
null
Predict
predictstories@gmail.com
predict
FUTURE,SINGULARITY,ARTIFICIAL INTELLIGENCE,ROBOTICS,CRYPTOCURRENCY
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Ankur Kashyap
Software Engineer, Seeker, Nature Lover.
9a5a5bd853da
ankurkashyap85
19
37
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-04-05
2018-04-05 17:19:45
2018-04-05
2018-04-05 17:21:46
0
false
en
2018-06-30
2018-06-30 03:35:39
7
1350b061caf6
2.871698
0
0
0
It’s difficult to fathom that just 20 short years ago — heck, even 10 years ago, or five — the technology we experience now in our everyday…
5
5 of the Fastest Growing Careers in 2018 It’s difficult to fathom that just 20 short years ago — heck, even 10 years ago, or five — the technology we experience now in our everyday lives was a mere figment of our imagination. Careers in technology were few and far between, or the ones that did exist required an excessive amount of schooling and experience. Nowadays, careers in technology are some of the fastest growing in their fields. It doesn’t even matter which field you’re in — academia, industry, commerce, etc. — tech jobs are expanding in each and every one. Together they usher in the new digital era and revolutionize the way we live and interact with each other in the new technological world. According to LinkedIn, “The tech boom and the rise of artificial intelligence are driving demand” for job growth. From 2012 to 2017, LinkedIn reported that five of the 20 fastest growing jobs in the U.S. are in development. Needless to say, coding plays a significant role in all of these career fields. Let’s take a look at five of the fastest growing careers in development. Machine Learning Engineer Forbes states, “With backgrounds and skills in data science, applied research and heavy-duty coding, [machine learning engineers] run the operations of a machine learning project and are responsible for managing the infrastructure and data pipelines needed to bring code to production.” Machine learning engineers straddle the line between knowing mathematics and coding the mathematics. As you might imagine, machine learning engineer combines both software engineering and data science experience — and it tops LinkedIn’s list of 20 fastest-growing careers. If you’re interested in coding, this field might just be the way to go. Data Scientist Data scientists tend to be part mathematician, part computer scientist and part trend-spotter. Because they have one foot in the business world and the other in IT, they’re highly sought-after and well-paid. Not to mention data scientist comes in right after machine learning engineer as number two on LinkedIn’s list. While data scientists weren’t particularly popular just a few years ago, their necessity has exploded as more businesses now consider big data. Data scientists mine through the unstructured information presented by businesses and develop key insights and analytics from it. Plus, the field of data scientists requires just a hint of programming and coding knowledge. Big Data Developer Coming in at number five on LinkedIn’s top twenty list is the big data developer. As Quora says, “The big data engineer builds what the big data solutions architect has designed. Big data engineers develop, maintain, test and evaluate big data solutions within organizations.” A big data developer integrates a proven comprehension of big data applications, solutions and programming. Quora goes on to say, “The big data engineer is a technical job that requires substantial expertise in a broad range of software development and programming fields.” For coding experts, the big data developer is a valuable career option. Unity Developer Developed by Unity Technologies, Unity is a cross-platform game engine, primarily used to develop both three-dimensional and two-dimensional video games and simulations for computers, consoles and mobile devices. Arguably the most popular game development software platform, there are currently 4.5 million registered Unity3D developers. Shortly following the big data developer on LinkedIn’s top 20 list is the Unity developer. Unity developers use programming to develop games, technologies and experiences in the emerging world of immersive and interactive media. Full Stack Developer Finally, rounding out the top 10 of LinkedIn’s fastest-growing jobs in the U.S. is the full stack developer. A full stack developer tends to be a jack of all trades. As Codeup says, “Full-stack die-hards would consider a full-stack developer to have specialized knowledge in all stages of software development.” This includes things such as server, network and hosting environment; user interface and user experience; and quality assurance. Coderbyte states full stack web development is the most popular developer occupation today. If you want to use coding to become a developer, this might be the perfect job for you. There’s no doubt about it: technology jobs are quickly taking over the world. If you want a job in tech, check out these five growing careers that utilize programming, data and development. Why Learn to Code at Midland University? There are 135 years of higher education experience behind this unique in-person program. In just 9 months, the Code Academy teaches you the skills needed to become a development professional. Learn more here.
5 of the Fastest Growing Careers in 2018
0
5-of-the-fastest-growing-careers-in-2018-1350b061caf6
2018-10-27
2018-10-27 20:02:42
https://medium.com/s/story/5-of-the-fastest-growing-careers-in-2018-1350b061caf6
false
761
null
null
null
null
null
null
null
null
null
Data Science
data-science
Data Science
33,617
Micah Yost
If opening too many Chrome tabs is talent, I’m Kanye. Amateur Investor. Founder & President at Method Mark, a brand strategy and marketing company.
ec2848c0389
micahyost
647
972
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-01-07
2018-01-07 18:53:58
2018-01-07
2018-01-07 18:53:55
8
false
en
2018-01-07
2018-01-07 18:53:59
5
135361ab8d05
3.933333
1
0
0
null
5
CoreML — iOS App Implementation for the Boston Price Model (Part 1) Hey! How are things? I hope the beginning of the year is looking great for you all. As promised, I am back to continue the open notebook for the implementation of a Core ML model in a simple iOS app. In one of the previous post we created a linear regression model to predict prices for Boston properties (1970 prices that is!) based on two inputs: the crime rate per capita in the area and the average number of rooms in the property. Also, we saw (in a different post) the way in which Core ML implements the properties of the model to be used in an iOS app to carry out the prediction on device! In this post we will start building the iOS app that will use the model to enable our users to generate a prediction based on input values for the parameters used in the model. Our aim is to build a simple interface where the user enters the values and the predicted price is shown. Something like the following screenshot: You will need to have access to a Mac with the latest version Xcode. At the time of writing I am using Xcode 9.2. We will cover the development of the app, but not so much the deployment (we may do so in case people make it known to me that there is interest). In Xcode we will select the “Create New Project” and in the next dialogue box, from the menu at the top make sure that you select “iOS” and from the options shown, please select the “Single View App” option and then click the “Next” button. This will create an iOS app with a single page. If you need more pages/views, this is still a good place to start, as you can add further “View Controllers” while you develop the app. Right, so in the next dialogue box Xcode will be asking for options to create the new project. Give your project a name, something that makes it easier to elucidate what your project is about. In this case I am calling the project “BostonPricer”. You can also provide the name of a team (team of developers contributing to your app for instance) as well as an organisation name and identifier. In our case these are not that important and you can enter any suitable values you desire. Please note that this becomes more important in case you are planning to send your app for approval to Apple. Anyway, make sure that you select “Swift” as the programming language and we are leaving the option boxes for “Use Core Data”, “Include Unit Tests” and “Include UI Tests” unticked. I am redacting some values below: On the left-hand side menu, click on the “Main.storyboard”. This is the main view that our users will see and interact with. It is here where we will create the design, look-and-feel and interactions in our app. We will start placing a few objects in our app, some of them will be used simple to display text (labels and information), whereas others will be used to create interactions, in particular to select input values and to generate the prediction. To do that we will use the “Object Library”. In the current window of Xcode, on the bottom-right corner you will see an icon that looks like a little square inside a circle; this is the “Show the Object Library” icon. When you select it, at the bottom of the area you will see a search bar. There you will look for the following objects: Label Picker View Button You will need three labels, one picker and one button. You can drag each of the elements from the “Object Library” results shown and into the story board. You can edit the text for the labels and the button by double clicking on them. Do not worry about the text shown for the picker; we will these with these values in future posts. Arrange the elements as shown in the screenshot below: OK, so far so good. In the next few posts we will start creating the functionality for each of these elements and implement the prediction generated by the model we have developed. Keep in touch. You can look at the code (in development) in my github site here. Originally published at Quantum Tunnel Website.
CoreML — iOS App Implementation for the Boston Price Model (Part 1)
1
coreml-ios-app-implementation-for-the-boston-price-model-part-1-135361ab8d05
2018-04-12
2018-04-12 02:58:59
https://medium.com/s/story/coreml-ios-app-implementation-for-the-boston-price-model-part-1-135361ab8d05
false
742
null
null
null
null
null
null
null
null
null
Apple
apple
Apple
44,415
Dr J Rogel-Salazar
Physicist... data science, quantum mechanics, ultra-cold matter, nonlinear optics, computational physics... not enough space.
588aafda9118
quantum_tunnel
323
461
20,181,104
null
null
null
null
null
null
0
null
0
721b17443fd5
2018-08-13
2018-08-13 16:47:07
2018-08-13
2018-08-13 19:20:35
12
false
en
2018-09-13
2018-09-13 16:24:26
0
1355acbe6f16
5.387736
4
0
1
Let’s take this Quiz and try to answer:
5
Machine Learning Simplified for Dummies Let’s take this Quiz and try to answer: (Note: try to guess the answer before continuing reading…) 3 -> 9 4 -> 16 7 -> 49 8 -> 64 10 -> ?? Did you answer 100? Why? Why not 46 or 67 or some other four-digit number or a letter from the alphabet? It is because you figured out that in the first four pairs of numbers the second number in each pair is the square of the first number. Analysis by the Human Brain Now, let’s say what if we give the same five pairs to a machine like the computer. Just as the human brain analyses the first four examples and prompts the result for the fifth one, we need machines to do the same. This is Machine Learning. Wherein, we want the machines to process the given input data (in our case, the first four examples) and for any new data (the fifth pair) guess or predict the outcome based on the previous input data that it processed. We want the machine to learn from experience and make appropriate guesses or predictions for new data points. In the above case it was just text — to be precise numbers. The input data could be anything ranging from text or numeric data to images, videos, audio’s, etc. And it is not only limited to the above types of data. Applications / Examples of Machine Learning: Analyzing Customer Behavior We want machines to predict, what a consumer is likely to buy from Walmart or Sam’s, considering his buying history for the past five years. Where Walmart has your data for past five years and knows every time you come to buy something, you definitely buy detergent, cooking oil, bread, snacks, chips, chocolates, etc along with the exact quantity. What if next time you go and Walmart already has this ready for you??? Or gets it delivered at your home. How did it know you exactly needed this? This is because the computer at Walmart that processes your bills knows what you buy every time and has now learned based on your shopping habits. Similarly, we already have Netflix and YouTube recommending videos to their subscribers based on their browsing history. This is Machine Learning. Google Photos — Identifying People by Face Recognition We might want computers to feed with an exhaustive number of pictures of a particular person A and then for any new human picture the computer should identify whether the new picture belongs to the person A or not. This is a real time example that Google uses for Google Photos. Similarly, voice over audio files of a person where for a given new audio file the computer predicts whether the voice belongs to the same person or not. We want software to classify emails as spam or not before reaching our mailboxes based on what emails the user treated as junk previously. In addition, we have applications to predict the weather forecast for the coming days, based on data regarding the temperature’s for past several hundred years, the timezone’s, the locations, the time of the year, etc. In definition by Arthur Samuel– Machine Learning is the ability of computers or machines to learn without being explicitly programmed. The machines could learn from the input data and for new data points give appropriate guesses based on input data without the need of programming for the new data point. Another definition by Tom Michel goes like this — A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E.” Predicting House prices To understand the definition better — let’s take an example of say, predicting house prices in the city of Dallas. Let’s say we fed a million data points to the computer program that consists of Dallas house rates including data per sq. feet, neighborhood area, etc., from the year 2005 to date. Here, E = a million data points T = prediction of house rate P = the difference between the predicted value and the actual value of the vendor. Lesser the difference, more accurate the prediction would be. Why is Machine Learning important today? Machine Learning is evolving. We deal with data through our daily chores whether at home or at school or at workplace. Data is the lifeblood of all businesses. Machine learning will enable us to make decisions in order to keep companies ahead of competition rather than falling behind. Two major advantages of machine learning is saving time and the aid of making appropriate decisions. Which University???? Damn…I need help!!! Let’s say a student is seeking to pursue a higher degree in a specific country. If data is available, a machine learning model could analyze the data of all those universities in that country and the student’s data and predict which university was the closest and best fit for the student in terms of ranking, location, fees, extra-curricular etc. Let’s say I develop this model and expand it for all countries in the world and host it as my product on a website. Now, students from all over the world could come to my website and see what the best university for them is based on their academic data, their preferred location and courses of their choice. Machine learning will generate revenue saving time for the customers and helping them in making appropriate decisions. Or the current YouTube recommendations. After viewing a particular video, based on the genre of the video, your location, the language, YouTube recommends / predicts other videos that you might like. It measures its performance based on whether you hit those recommended videos or not. Thus saving your time where you do not need to browse for other videos of the same type. This is so efficient when you are looking to learn something new at the last minute. The predictions would give you a collection of all the videos related to the topic you are searching for. Thus saving your time, helping you in getting information from the best videos and in turn increasing the YouTube’s business. This is simply Machine Learning. Now, you might be thinking how can a machine predict what a consumer is likely to buy, or which video a consumer is likely to look for. That is an entire different world which includes a combination of statistics, algorithms, data analysis to name a few.
Machine Learning Simplified for Dummies
90
machine-learning-simplified-for-dummies-1355acbe6f16
2018-09-13
2018-09-13 16:24:26
https://medium.com/s/story/machine-learning-simplified-for-dummies-1355acbe6f16
false
1,070
Coinmonks is a technology focused publication embracing all technologies which have powers to shape our future. Education is our core value. Learn, Build and thrive.
null
coinmonks
null
Coinmonks
gaurav@coinmonks.com
coinmonks
BITCOIN,TECHNOLOGY,CRYPTOCURRENCY,BLOCKCHAIN,PROGRAMMING
coinmonks
Machine Learning
machine-learning
Machine Learning
51,320
Caston Fernandes
Data Scientist in-the-making!
702a524e8a13
casternxavier
7
12
20,181,104
null
null
null
null
null
null
0
null
0
7fbce5859bab
2017-10-25
2017-10-25 08:38:42
2017-10-25
2017-10-25 19:50:24
2
false
en
2017-11-15
2017-11-15 12:06:13
2
1355d1a2c84d
1.37956
0
0
0
Ever wondered what it takes to get a song at the top? What makes Kanye, Mariah Carey, Bee Gees, Michael Jackson and The Beatles so easy to…
5
Will that song your band just wrote be a hit? We’ve found the tech that’ll answer the question. Ever wondered what it takes to get a song at the top? What makes Kanye, Mariah Carey, Bee Gees, Michael Jackson and The Beatles so easy to listen to? Well, there’s a project that strives to answer all of these questions — thanks to Artificial Intelligence. We’re of course talking about Hitwizard, AI created by Goldmund Wyldebeast & Wunderliebe, which uses self-learning algorithms. You’re probably already familiar with the mechanism — Facebook and Google use them to determine what’s in a photo. The same techniques are now being applied to songs — with the AI learning what makes a song succesful (like BPM, tempo, chord progressions and sequences). Currently, Hitwizard is actually better at predicting if a song will fail: “From 85,000 different tracks that have been played on the Dutch radio during the last six months, nearly 60% of the predictions of the Hitwizard for a song to become a hit were correct. As opposed, nearly 99.6% of cases classified by the Hitwizard as non-hits were predicted correctly as well”, stated the lead developer of the Hitwizard. Ready to do some estimations? Head to Hitwizard’s website and try it out for yourself. 🎶 Aaand if you’re into music… we’re Amondo, a startup that wants to help you capture your best gig experiences by pulling together the best photos & videos from the night — from you, your friends, other fans, even the artist on stage. Check our website out for more info: amondo.com
Will that song your band just wrote be a hit? We’ve found the tech that’ll answer the question.
0
will-that-song-your-band-just-wrote-be-a-hit-weve-found-the-tech-that-ll-answer-the-question-1355d1a2c84d
2018-02-23
2018-02-23 11:04:17
https://medium.com/s/story/will-that-song-your-band-just-wrote-be-a-hit-weve-found-the-tech-that-ll-answer-the-question-1355d1a2c84d
false
264
The Experience Platform. ⚡
null
helloamondo
null
Amondo
hello@amondo.com
amondo
AMONDO,STARTUP,APP DEVELOPMENT,CONTENT CURATION,LONDON
helloamondo
Music
music
Music
174,961
MJ Widomska
YRSTRULY.UK
17ad0df5a173
mjwidomska
26
12
20,181,104
null
null
null
null
null
null
0
null
0
3fb317be81a5
2018-08-30
2018-08-30 12:30:39
2018-08-30
2018-08-30 12:34:23
1
false
ko
2018-08-30
2018-08-30 12:35:25
0
1358ef6b5322
3.912
0
0
0
[MLY 번역]
3
[MLY 번역] 11. When to change dev/test sets and metrics [MLY 번역] Andrew ng의 Machine Learning Yearning을 번역한 글입니다. 중간중간 실험가능한 내용이 있을 경우 c++ 혹은 tensorflow를 이용한 실험과 함께 업로드할 예정입니다. 일주일에 2회, 회당 5개씩 업로드될 예정입니다. 번역: 안현진,정주안 새로운 프로젝트를 시작하려 할때, 저는 디벨롭셋과 테스트셋을 빠르게 선택하려고 노력합니다. 왜냐하면 이것이 팀원에게 잘 정의된 목표를 제공하기 때문입니다. 저는 일반적으로 제 팀에게 초기 디벨롭셋/테스트셋과 지수를 일주일 안에 만들 것을 요청합니다-가끔은 좀 더 걸립니다. 지나치게 생각만 하는것 보다는, 불완전한 것을 생각해내고 빠르게 나아가는게 더 낫습니다. 그러나 이런 한주 타임라인은 발달된 분야(mature applications)에 적용하기에 부적합합니다. 예를들어 안티-스팸(anti-spam)은 이미 많이 발달된 딥러닝 분야입니다. 저는 이미 성숙한 시스템에서 더 나은 디벨롭셋과 테스트셋을 얻기 위해 몇달을 작업하는 팀의 모습을 보아 왔습니다. 만약 초기 디벨롭셋과 테스트셋 또는 지수가 주요한 점를 놓친것을 나중에 알게된다면, 부디 신속하게 변경하시기 바랍니다. 예를 들어 만약, 디벨롭셋과 지수가 A분류기를 B분류기보다 높게 평가했더라도 여러분의 제품에 B분류기를 적용하는게 더 우수하다고 판단된다면 이건 아마도 여러분의 디벨롭셋/테스트셋 혹은 평가지수를 바꿔야 함을 알리는 것일 수도 있습니다. 분류기 A가 더 좋다고 잘못된 평가를 내리게 된 세가지 가능한 주된 요인은 아래와 같습니다. 실제 데이터셋의 분포가 디벨롭셋/테스트셋의 분포와 달랐을 경우 입니다. 여러분의 초기 디벨롭셋과 테스트셋이 주로 다 큰 고양이의 사진으로 구성되었다고 가정해봅시다. 여러분의 고양이앱이 배포되고, 유저들은 여러분의 기대와 달리 다 큰 고양이보다 새끼고양이의 사진을 올렸고 따라서, 여러분의 디벨롭셋과 테스트셋은 분포는 실제 데이터의 분포를 대표할 수 없게 됐습니다. 이런 경우 더 잘 대표할 수 있는 값으로 디벨롭셋과 테스트셋을 업데이트 해주어야 합니다. 2. 디벨롭셋에 대해 과적합이 일어났을 경우 입니다. 디벨롭셋에 대한 아이디어를 반복적으로 평가하는 과정은 여러분의 알고리즘이 디벨롭셋에 대해 점차 과적합 되도록 만듭니다. 개발이 완료되면 여러분은 테스트셋에 대한 평가를 할 것입니다. 만약 디벨롭셋에 대한 퍼포먼스가 테스트셋에 대한 것보다 훨씬 높게 나온다면, 디벨롭셋에 대해 과적합이 일어났다는 신호일 수 있습니다. 이 경우, 새로운 디벨롭셋을 얻어야 합니다. 만약 팀의 진행상황을 추적해야하는 경우, 테스트셋에 대한 실험을 일주일 혹은 한달에 한번 시행하여 시스템을 정기적으로 평가할 수도 있습니다. 그러나 테스트셋을 사용하여 이전 주의 시스템으로 돌리는것을 포함한 알고리즘과 관련된 어떤 결정도 하지 마십시오. 만일 그리한다면 시스템은 테스트셋에 대해 과적합될 것이며 시스템의 퍼포먼스에 대한 공정한 평가치를 제공하지 않을 것입니다.(공정한 평가치는 연구논문을 퍼블리싱하거나 업무에서의 중요한 의사결정을 할 때 필요할 것입니다.) 3. 지수가 프로젝트가 최적화하고자 하는 문제가 아닌 다른 것을 측정할 때입니다. 고양이 애플리케이션을 생각해봤을때, 여러분의 지수는 분류의 정확성(accuracy)입니다. 이 지수는 현재 분류기 A가 B보다 우수하다고 평가합니다. 그러나 두가지의 알고리즘을 다 실험해보고, 분류기A가 가끔 포르노 이미지를 통과시킬 수 있다고 생각해보십시오. 분류기 A의 정확성이 더 높다고 하더라도, 때때로 통과되는 포르노 이미지 때문에 나쁜인상을 남길 수 있다면, 그것의 퍼포먼스가 용납될 수 없음을 의미합니다. 이런 경우 당신은 어떻게 하시겠습니까? 여기서, 그 지수는 여러분의 제품에 대해 알고리즘 B가 실제로 알고리즘 A보다 더 뛰어나다는 사실을 식별하지 못합니다. 따라서 더 이상 최상의 알고리즘을 선정하기 위해 그 지수를 신뢰할 수 없습니다. 평가지수를 바꿔야할 때 입니다. 예를 들면 포르노 이미지에 대해 무거운 처벌을 주도록 지수를 변경할 수 있습니다. 저는 신뢰할 수 있는 지수 없이 너무 오랫동안 진행하고 수동으로 분류자를 선택하는 것으로 되돌아가는 것보다, 새로운 지수를 뽑고 이를 이용해 팀의 명확한 목표를 정의하는것을 강력 추천합니다. 프로젝트 진행 중 디벨롭셋/테스트셋이나 평가지수를 변경하는 것은 아주 흔한 일입니다. 초기에 설정한 디벨롭셋/테스트셋과 지수는 여러분의 반복과정을 빠르게 하는데 도움을 줍니다. 여러분이 디벨롭셋/테스트셋이나 지수가 더이상 옳지 않다고 판단되었을때, 이것은 큰 문제가 되지 않습니다. 단지 새로 설정하고 여러분의 팀이 새로운 방향을 확실하게 알도록 하면 됩니다.
[MLY 번역] 11. When to change dev/test sets and metrics
0
mly-번역-11-when-to-change-dev-test-sets-and-metrics-1358ef6b5322
2018-08-30
2018-08-30 12:35:25
https://medium.com/s/story/mly-번역-11-when-to-change-dev-test-sets-and-metrics-1358ef6b5322
false
520
‘Machine Learning Yearning’ written by Andrew ng, translated to korean.
null
null
null
MLY_KOREA
johnnyapu15@gmail.com
mly-korea
MLY,MACHINE LEARNING,DEEP LEARNING
null
Machine Learning
machine-learning
Machine Learning
51,320
안현진
null
2ea55c99eda
hjan7804
2
2
20,181,104
null
null
null
null
null
null
0
null
0
8ef729cf5aef
2018-03-07
2018-03-07 11:29:20
2018-03-07
2018-03-07 11:34:25
7
false
en
2018-03-07
2018-03-07 11:34:25
1
135ad677c5c5
4.178302
1
0
0
Originally published on The Tech Brewery
5
Top Technology Trends In 2018 Originally published on The Tech Brewery It is 2018 already and there is nothing that is impossible in today’s world. Why? Because we have the superpower called technology. We, humans have become dependent on technology for everything. The smartphones have invaded our lives completely and there are other technologies, still in development that will disrupt our lives once again. While it is the most difficult task to predict what technology has to offer in future, this blog is based on the big trends that we are expecting in the year ahead. Artificial Intelligence AI(Artificial Intelligence) is one of the growing revolutionary technology in the business world as well as in science. Nowadays many IT companies are investing in AI. Big giants tech companies like Google and Baidu investing around $20 — $30 Billion on AI in 2016. Artificial Intelligence is basically a machine which has the ability to copy intelligent behaviour. Artificial Intelligence possibilities are endless, it is already present in our phones and we can find this with the help of Mobile, it can be embedded using chatbots or context aware sensors. When we are thinking about the smart applications or artificial intelligence (AI) Siri, Cortana, or Assistant probably come to our mind. Hundreds of linguists and software engineers dedicate countless hours to building these services into responsive personal assistants that can answer questions, track down information, send messages, launch services, and more. Virtual Reality In layman language, “Virtual” means artificial or digital and “Reality” is what we experience as a human. Virtual reality makes possible to experience anything, anytime, anywhere. Virtual Reality provide us an artificial environment which is experienced through sights and sounds which itself provided by computer and our actions determine what happens in the environment. As a kid, we were told about the five senses that we possess (taste, touch, smell, sight, and hearing). However, neurologists have identified 21 senses and with these senses we experience the virtual environment around us. VR Games are becoming popular among masses. A lot of industries are ready to experiment with this upcoming technology to offer amazing yet virtual experience to the target audience. Augmented Reality If you have used Snapchat, well you are a user of Augmented Reality already. Augmented Reality, in simple terms is when you add some virtual effects to the real-time scenario of the user. Apple has introduced the animated emoji that use augmented reality. Last year, PokemonGo was one of most popular players in this domain. Majority of the people might think VR and AR are the same things but in reality these are completely different concepts. AR enhances your reality with virtual effects whereas VR takes you to another world by taking you away from reality with the help of some gears. Cryptocurrency Cryptocurrency is the latest buzz word. You might have heard someone or the other talking about investing in some kind of cryptocurrency. While Bitcoin was the first cryptocurrency that got popular across the globe, there is a huge list of different cryptocurrencies that you still have not heard of. Cryptocurrencies are some database entries that cannot be changed by anyone without accomplishing certain specific conditions. Cryptocurrency market is hot and trending everyday with different new additions to it. There are exchanges that enable trading of the cryptocurrencies. A lot of new cryptocurrencies come and go in the market. A few of them turn out to be promising while others fail in the first few months leaving investors with abandoned currency. However, people are still accepting it at a large scale because there will be no manipulation on the monetary front by any bank or government. Huge investments are being made on cryptocurrencies as people believe that this will change the world. Internet Of Things Originally published on The Tech Brewry ‘Internet’ of things refers to a network created by several devices with sensors, enabling the seamless sharing of information through bare objects. Things like smart mirrors, clothing, umbrellas and what-have-you are flooding the market, and they are all part of the new IoT revolution. In effect, the concept details that everything can be interconnected. You might have heard about Smart Homes. Well, that is entirely based on the concept of IOT. Connected gadgets let you manage all your home activities through your smartphone and optimize your daily tasks and streamline operations efficiently. Big Data Data is the basic entity that is being generated enormously via millions of sources and is then distributed across the sea of networks. This implies that huge amount of data is being transferred over the internet. To help organizations manage the useful and authentic data, Big Data is the ultimate solution. Enterprises are becoming more data-driven which directly increases the demand and importance of Big Data. Business insights and relevant numbers help organizations frame their marketing strategies as well. Explore what all you can do with these technology trends or contact us for a free consultation.
Top Technology Trends In 2018
1
top-technology-trends-in-2018-135ad677c5c5
2018-03-09
2018-03-09 03:49:17
https://medium.com/s/story/top-technology-trends-in-2018-135ad677c5c5
false
829
Code Brew Labs brings the most happening ideas to reality. With the best team of developers and designers, we have over 300 applications live on Google Play Store and App Store.
null
codebrewlabs
null
code-brew.com
business@code-brew.com
code-brew-com
STARTUP,APP DEVELOPMENT,APPS,INFORMATION TECHNOLOGY,ENTREPRENEURSHIP
codebrewlabs
Virtual Reality
virtual-reality
Virtual Reality
30,193
Sakshi Ghavri
null
1e3d5bce3084
sakshighavri
14
26
20,181,104
null
null
null
null
null
null
0
python3 -m pip install --upgrade pip python3 -m pip install jupyter jupyter notebook
2
null
2018-09-18
2018-09-18 11:31:51
2018-09-20
2018-09-20 11:41:21
8
false
en
2018-09-20
2018-09-20 11:41:21
5
135b53bc7168
3.850314
1
0
0
If you are an absolute beginner, and want to know how to get everything you need on your computer, you can read on.
5
Setting up data science work space for beginners (Jupyter Notebook) If you are an absolute beginner, and want to know how to get everything you need on your computer, you can read on. The two most commonly used tools for data science are python and R. There are a lot of IDEs available for data science projects such as Spyder, PyCharm, R Studio, Rodeo, Atom, Sublime Text etc. But the one tool which every data scientist will be comfortable to work with is Jupyter Notebook(previously known as Ipython). The Jupyter Notebook App is a server-client application that allows editing and running notebook documents via a web browser. It can be accessed without internet access as well. Notebook documents (or “notebooks”, all lower case) are documents produced by the Jupyter Notebook App, which contain both computer code (e.g. python) and rich text elements (paragraph, equations, figures, links, etc…). Installing Jupyter using Anaconda(recommended) Anaconda is a free and open source distribution which includes data science packages for Python/R and manage your packages, dependencies, and environments. You can download the latest version of anaconda here. Install Jupyter Notebook using pip If you already have Python 3 installed (which is recommended): To run the notebook, run the following command at the Terminal (Mac/Linux) or Command Prompt (Windows): JupyterLab JupyterLab is considered as evolution of Jupyter Notebooks. JupyterLab is just like jupyter notebook but is much more productive by enabling you to work with documents and activities such as Jupyter notebooks, text editors, terminals, and custom components in a flexible, integrated, and extensible manner. Create your first notebook Once you have installed Aanconda, you can open the anaconda navigator and launch Jupyter Notebook. Now that you have your jupyter notebook app running, you can open an existing file to work with or create a new one. It has four options to choose from i.e., Text File gives you a blank text editor. Folder allows you to create new folder and add files in it. Terminal is exactly like the command prompt in Mac, Linux(cmd in Windows). Python 3 allows you to run python code, describe the code with Markdown support. If you have already seen the notebooks published by other educators or professionals, you may have observed that they included texts, links, images between the code. This is possible because of the Markdown feature. Image showing Markdown, Code and output below the code cell The cells in notebook can be any of these four types i.e., Code, Markdown, Raw NBConvert, Heading. Since Anaconda is installed along with python and its libraries and packages, you can start writing your code. Write your python code in the empty cell. To run the code cell, all you need to do is press Ctrl+Enter . Now, it not only runs your code but also shows you the results immediately. When you press Ctrl+Enter , the cell shows [*] in the brackets, which once finished is replaced with the number [1], showing the start of the session. You can see the output just below the cell. For more cells, you can insert the cell by clicking the ‘+’ button or click Shift+Enter while running code. So it automatically inserts a new cell below along with the output. Or you can just click Insert from the menu and select one of the two options. Markdown You can use command mode (M) or just select from the drop down to change the cell to Markdown . To change the size of markdown text, use #. Image showing Markdown, Heading, Code and output below the code cell Magic functions Magic functions are built-in functions which allow you to work much like OS command-line calls.There are two types of magics. Line Oriented prefixed by % ,for single line input. Cell Oriented prefixed by %% ,for multiple line input. A simple magic function which is very commonly used is %matplotlib inline which allows you to show the graph plotted in the notebook itself. If you want to know other existing magics, run %lsmagic, you’ll get the list of all the functions. To know more about magics, run %magic which provides the document about IPython’s ‘magic’ functions. You can even create custom magics by inheriting from a base class provided by IPython: IPython.core.magic.Magics. For more information you can check the documentation here Thank you for reading. Hope you learnt something new:)
Setting up data science work space for beginners (Jupyter Notebook)
5
setting-up-data-science-work-space-for-beginners-jupyter-notebook-135b53bc7168
2018-09-20
2018-09-20 11:41:21
https://medium.com/s/story/setting-up-data-science-work-space-for-beginners-jupyter-notebook-135b53bc7168
false
720
null
null
null
null
null
null
null
null
null
Data Science
data-science
Data Science
33,617
Aakanksha
null
2131ed332b15
aakankshaws
1
21
20,181,104
null
null
null
null
null
null
0
null
0
48cc3c93dad0
2018-02-07
2018-02-07 20:36:59
2018-02-07
2018-02-07 20:40:22
1
false
en
2018-04-11
2018-04-11 18:28:12
1
135bbb4bb781
3.407547
4
0
0
Deepening Consumer-brand Relationships
4
Marketers as Relationship Scientists Deepening Consumer-brand Relationships Advertising as we know it is dying — and that is a good thing. From basic technical innovation (e.g., ad blocking going “native” in browsers,) to social trends (like cord-cutting,) to new core values (like Gen Z’s caring less and less about brands,) everything points in a similar direction. More people than ever are trying to bypass advertising that only asks from them without providing much value. It’s the end of interruptions. An Age of Relationship Scientists Who will benefit from all of this? On one end, consumers will. Despite the omnipresence of innovators like Amazon, consumers will continue to circumvent irrelevant interruptions with an ever diminishing effort as they go about their lives. So, who will benefit within our industry? We believe this to be an age of relationship science — where brands and marketers have a vast amount of data at their disposal, and if used correctly, could build genuine bonds to sustain their creativity. A New Consumer-brand Relationship Paradigm Strong relationships between brands and consumers under this context need to comply with three basic features: Unique: engagement at the individual level, in a specific, granular way. Meaningful: relevant and impactful interactions between brands and consumers. Self-improving: actions tied to business results, short and long-term, on and offline. Let’s dissect these three pillars. Unique Many brands try too hard to be unique, becoming cliché (“you should listen to what I’m saying, because it’s ‘unique.’”) Here we are referring to something completely different. Successful brands, including some we work with, see growth by engaging with consumers individually, and this is what unique should mean. When we say ‘truly individual’ it’s not a metaphor. One by one, like the relationships you have with each person in your life, brands are forging real bonds. Scaling this to the hundreds of thousands or millions is not easy, but it’s never been more attainable. As advertisers, many of us use persona-building as a tactic to learn more about our audience and how to engage with them. Still, personas are only useful as a starting point. People-based marketing, and the models within it, are changing the advertising landscape forever. On top of these basic archetypes, we want to ensure every piece of available data is used to establish concrete relationships at a one-to-one level. Meaningful As marketers, we should aim for our brands to be discovered, not pushed into people’s minds. We don’t want our brands to be the sign at the side of the road — we want them to be the destination. As much as impactful, creative storytelling will help us break through the clutter to reach our target audiences, this should be considered solely as the price of admission; that is not a relationship yet. In order to build progressively deeper bonds, brands need to establish a permanent flow of micro-impacts at the individual level. This is best expressed with a simple motto: from 10 to 10,000. Evolving from the 10 big executions a brand would use to engage with their consumers on any given year in the past, we now need and want to multiply that by 1,000. These 10,000 smaller scale executions will help us impact our relationships with consumers in more intimate and meaningful ways. A solid “micro-persona” ecosystem — beyond generic consumer definitions, paired with people-based marketing — helps us build stronger and deeper relationships based on relevance and timing. What brands choose to do with these deeper relationships should be carefully aligned with the trust that has been created. Self-improving It’s important not to fall in the trap of oversimplifying what a self-improving brand relationship is. As an industry, we claim “everything is measurable;” it can be easy to set our KPIs and try to track the depth of our relationships through them. At the same time, for these bonds with individual consumers to grow deeper and stronger, every marketer needs to act as a relationship scientist. We need to stop thinking of data as a matter only geeks deal with, and allow ourselves to play with the numbers. To strengthen our relationships with consumers we must leverage 1st and 3rd party data, tie online engagement with physical store traffic, and connect CRM segments with social graphs. Intentional, thoughtful and sensitive usage of data means you care about both your consumer and your brand. From assessing the impact of a channel on another, to understanding how a unique relationship between a consumer and a brand are born, connections can either grow stronger or apart. As an industry, we need to evolve (and are currently evolving) the conversation from mere attribution modeling to fully individual consumer journeys. Originally published in MediaPost on January 5, 2018. Jonatan Zinger, SVP, Media Insights — M8 With 17+ years of experience in the digital advertising industry, Jonatan’s profile combines online media and a strategic background focused in analytics. He’s led online marketing efforts for Fortune 500 brands in travel, entertainment, electronics, CPG, business services, education and automotive, among other verticals.
Marketers as Relationship Scientists
17
marketers-as-relationship-scientists-135bbb4bb781
2018-04-11
2018-04-11 18:28:13
https://medium.com/s/story/marketers-as-relationship-scientists-135bbb4bb781
false
850
An agency that crafts meaningful relationships between brands and consumers in a Connected Life
null
M8agency
null
M8 Connected Life
john.santiago@M8agency.com
m8connectedlife
ADVERTISING,MARKETING,DIGITAL MARKETING,MULTICULTURAL
M8agency
Marketing
marketing
Marketing
170,910
Jonatan Zinger
SVP, Media Insights, M8
ae65c9ce439d
JonZinger
52
46
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-03-21
2018-03-21 13:55:42
2018-03-21
2018-03-21 13:55:43
1
false
en
2018-03-21
2018-03-21 13:55:43
1
135c6596146
1.950943
2
0
1
null
5
What Does Big Data Mean in 2018? Big data was still big news for much of 2017, thanks in part to the accelerated growth of the Internet of Things (IoT). The proliferation of smart devices we saw being put to practical use last year — at both consumer and business level — meant we were constantly being reminded of the speed at which big data was, in layman’s terms, getting ‘bigger’. However, another trend emerged during 2017 that runs almost counter to the above: as this site noted in a round-up of emerging trends at the end of last year, the concept of big data itself arguablyceased to be news at all. Between 2017’s constant talk of smart home products, smart customer analytics and service, smart production line hardware and smart logistics solutions, we became inured to those once-newsworthy statistics about the sheer volume of information. Indeed, the idea of big data itself became almost mundane; most companies simply reverted to calling it ‘data’. It ceased to qualify as a buzz term. If increasingly widespread acceptance was a defining feature of 2017, we could expect 2018 to be about focused, real-world applications. In short, the impact of big data is set to become far more tangible over the coming months than ever before. So far, to the majority public at least, big data has tended to remain a somewhat woolly concept — a vague understanding that, due in part to the IoT, more and more of our daily interactions are being logged to colossal, ever-growing spreadsheets whose vast potential we haven’t yet figured out how to fully unlock. 2018, by contrast, is all about the real-world application of that information, and how it can make a practical difference to our everyday lives. In turn, we finally see the typical conversation around big data begin to shift from a spreadsheet-based, quantitative phenomenon to more of a qualitative real-world one. 2017 was a huge year for the concept of machine learning. Along with impressive leaps forward in the complex technology associated with teaching computers to interpret and react to emerging datasets independently, we also saw increased public concern about how exactly this might impact on the job market in future. However, things are looking far from bleak in that regard. While there will, of course, be some traditional, relatively basic jobs thateventually lose out to AI as it continues to infiltrate more of our routine daily interactions, we’re also likely to see a huge uptick in the number of positions becoming available because of big data. Indeed,IBM projections indicate that some 2.7m people will be employed indirectly data-related positions by 2020, some 700,000 of which will specifically be in advanced data science and analytics roles. Posted on 7wData.be.
What Does Big Data Mean in 2018?
4
what-does-big-data-mean-in-2018-135c6596146
2018-03-21
2018-03-21 21:50:44
https://medium.com/s/story/what-does-big-data-mean-in-2018-135c6596146
false
464
null
null
null
null
null
null
null
null
null
Big Data
big-data
Big Data
24,602
Yves Mulkers
BI And Data Architect enjoying Family, Social Influencer , love Music and DJ-ing, founder @7wData, content marketing and influencer marketing in the Data world
1335786e6357
YvesMulkers
17,594
8,294
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-04-26
2018-04-26 13:18:38
2018-04-26
2018-04-26 13:20:15
1
false
en
2018-04-26
2018-04-26 13:20:15
2
135e14447f16
2.660377
0
0
0
24 Apr, 2018
5
Machine Learning, the booming career option shaping the job market 24 Apr, 2018 You’d be hard-pressed to find a field generating more buzz than Machine Learning these days. Jobs related to Machine Learning are growing rapidly as companies try to get the most out of emerging technologies. The most compelling thing about Machine Learning is its seemingly limitless applicability. There are hardly any fields (virtually) where machine learning does not apply. The fields that are immensely impacted by machine learning include education, finance, computer science and more, healthcare etc. So, one can easily guess that several machine learning skills are in high demand in the global marketplace today. The Indian IT industry is expected to add around 1.80 lakh to 2 lakh new jobs in 2018. This year will see a sharp increase in demand for professionals with skills in emerging technologies such as Artificial Intelligence (AI) and machine learning, even as people with capabilities in Big Data and Analytics will continue to be the most sought after by companies across sectors, say sources in the recruitment industry. Now, you might be wondering what it takes to get a machine learning job. One must have a deep understanding of a broad set of algorithms and applied math, problem-solving and analytical skills, probability and statistics and programming languages such as Python, C++, Java etc. Above all, to explore this subject need innate curiosity. In addition to that, there are few skills required to learn the subject and get a job in this field: With programming languages like Python and R and their packages making it easy to work with data and models, it is reasonable to expect a data scientist or machine learning engineer to attain a high level of programming proficiency and understand the basics of system design. So, one must be able to write high-quality software. It is necessary to work with large data sets. As the cost of data storage is on a dramatic downward trend, there are more and more data sets from different domains to work with and apply models to. Try to expand your expertise in Unix Tools. Since all of the processing will most likely be on the Linux-based machine, you need to learn their functions to utilize them well. Problem-solving skills are required to learn and understand the signal processing techniques. Feature extraction is one of the most important parts of machine learning. Depending on the type of problem you are trying to solve, you may be able to utilize really cool advance signal processing algorithms like wavelets, shearlets, curvelets, contourlets, bandlets. Learning about time-frequency analysis, and applying it to your problems will also help. To perform quantitative analysis on the system, product intuition is a necessary skill. This includes understanding the complex system that generates all of the data that data scientists analyze. A data scientist who understands the product well can generate hypotheses about ways the system can behave if changed in a particular manner. The traditional analytics skill set includes defining key primary and secondary metrics that the company can use to keep track of success at particular objectives. Staying curious about your work surroundings and industry remains a critical quality in an employee or a learner and is always in demand. This shows that you are investing time and energy to update yourself on new concepts, ideas. Machine learning is already an incredibly powerful tool that can do a surprisingly good job of solving really difficult classification problems. The next tier of demand is for people who can build practical implementations, especially in collaboration with a cutting-edge research team. Simply put, the popularity of machine learning has grown with the expansion and accessibility of data analytics in various industries and it’s a major advancement in the way we dwell in our local and global economy. Machine Learning Using Python » Originally published at www.isoeh.com.
Machine Learning, the booming career option shaping the job market
0
machine-learning-the-booming-career-option-shaping-the-job-market-135e14447f16
2018-04-26
2018-04-26 13:20:16
https://medium.com/s/story/machine-learning-the-booming-career-option-shaping-the-job-market-135e14447f16
false
652
null
null
null
null
null
null
null
null
null
Data Science
data-science
Data Science
33,617
Alisha Roy
null
fd167261198e
alishaethhack
1
1
20,181,104
null
null
null
null
null
null
0
=WEEKNUM( serial_number, [return_type] ) =WEEKNUM(A1:A25+0) =SUMPRODUCT((WEEKNUM(Table13[Order Date]+0)=[@[Week No.]])*Table13[Sales]) {34;23;23;45;45;24;24;35;16;16;23;35;35;45;23;45;35;34;35} {TRUE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;TRUE;FALSE} {18.504;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;68.81;0} {TRUE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;TRUE;FALSE} {18.504;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;68.81;FALSE} =SUM({18.504;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;FALSE;68.81;FALSE})=87.314
9
7f60cf5620c9
2018-02-11
2018-02-11 03:23:19
2018-02-11
2018-02-11 03:54:48
4
false
en
2018-02-11
2018-02-11 03:54:48
3
136103855345
2.601887
1
0
0
Every organization requires weekly summaries to monitor peak and off-peak periods. This is mostly so when you are in the production…
5
4 WAYS TO SUM DATA BY WEEK NUMBER Every organization requires weekly summaries to monitor peak and off-peak periods. This is mostly so when you are in the production section. Excel has a beautiful but overlooked function (WEEKNUM) that returns an integer representing the week in the year (1 to 53). For anyone not familiar with WEEKNUM function, it contains 2 arguments; Serial_number ( the date to return the week number for) and an optional Return_type (an integer that defines when the week starts). NB: If return type is omitted, function defaults to week start on Sunday all the way to Saturday Beautiful as the function is, it has one major weakness, WEEKNUM doesn’t accept a range argument e.g. =WEEKNUM(K11:K110) just returns #VALUE! error. So, How do you get week numbers in an array given a range of dates? Do you need to always use helper column to convert dates to week numbers before any analysis is done? In this article, I will show you 3 ways to simplify your weekly analysis: SUMIF SUMPRODUCT SUM & IF PIVOT TABLES SUMIF Given below data, Show totals per week. Since WEEKNUM does not accept a range argument, then we have to create a helper column so that we can be able to use SUMIF SUMPRODUCT It is not entirely true that WEEKNUM does not accept a range argument. It can be forced to accept range by adding zero to a range i.e. The above function does not return an error but an array of week numbers from the given dates range. Since SUMPRODUCT function comfortably handles arrays, we can use this array of week numbers to create a summary as shown below. How it works: ►WEEKNUM(Table13[Order Date]+0) returns an array of week numbers for the given data range ►WEEKNUM(Table13[Order Date]+0)=[@[Week No.]] returns an array of TRUE/FALSE based on the week number criteria. ►(WEEKNUM(Table13[Order Date]+0)=[@[Week No.]])*Table13[Sales] SUMPRODUCT converts the array of TRUE/FALSE into its numeric equivalent of 1/0 when multiplying it with the sales figure =SUMPRODUCT({18.504;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;68.81;0})=87.31 SUM & IF FUNCTION This is an array function that uses the same principles as SUMPRODUCT {=SUM(IF(WEEKNUM(Table135[Order Date]+0)=[@[Week No.]],Table135[Sales]))} How it works: ►WEEKNUM(Table13[Order Date]+0)=[@[Week No.]] returns an array of TRUE/FALSE based on the week number criteria. ►IF function returns only the Sales values if the test is TRUE, otherwise returns FALSE ►Since SUM function ignores texts, It just sums up the numbers PIVOT TABLES This is the simplest and easiest method. Its only major drawback is that it will give you the weeks date range but not the week number. http://crispexcel.com/wp-content/uploads/2018/02/video.mp4 NB: The trick in using pivot table is knowing how to group the dates into weeks. Watch the video again. Conclusion There is now no need to lose your hair over worrying how to do weekly summaries. You can use the same techniques taught above to do Average and Count. If I have missed something, share. DOWNLOAD SPREADSHEET FOR PRACTICE Originally published at crispexcel.com.
4 WAYS TO SUM DATA BY WEEK NUMBER
1
4-ways-to-sum-data-by-week-number-136103855345
2018-07-10
2018-07-10 22:02:08
https://medium.com/s/story/4-ways-to-sum-data-by-week-number-136103855345
false
504
Sharing concepts, ideas, and codes.
towardsdatascience.com
towardsdatascience
null
Towards Data Science
null
towards-data-science
DATA SCIENCE,MACHINE LEARNING,ARTIFICIAL INTELLIGENCE,BIG DATA,ANALYTICS
TDataScience
Data Science
data-science
Data Science
33,617
Crisp o Mwangi
Excel Enthusiast | Blogger |Data Scientist
d9968d0b2c30
crispomwangi
6
9
20,181,104
null
null
null
null
null
null
0
scala> val statsm = missingStats(parsed.filter(_.matched).map(_.scores)) statsm: Array[NAStatCounter] = Array( stats: (count: 20922, mean: 0.997316, stdev: 0.036506, max: 1.000000, min: 0.000000) NaN: 9, stats: (count: 1333, mean: 0.989890, stdev: 0.082489, max: 1.000000, min: 0.000000) NaN: 19598, stats: (count: 20931, mean: 0.997015, stdev: 0.043118, max: 1.000000, min: 0.000000) NaN: 0, stats: (count: 475, mean: 0.969370, stdev: 0.153291, max: 1.000000, min: 0.000000) NaN: 20456, stats: (count: 20931, mean: 0.987292, stdev: 0.112013, max: 1.000000, min: 0.000000) NaN: 0, stats: (count: 20925, mean: 0.997085, stdev: 0.053914, max: 1.000000, min: 0.000000) NaN: 6, stats: (count: 20925, mean: 0.997945, stdev: 0.045285, max: 1.000000, min: 0.000000) NaN: 6, stats: (count: 20925, mean: 0.996129, stdev: 0.062097, max: 1.000000, min: 0.000000) NaN: 6, stats: (cou... scala> val statsn = missingStats(parsed.filter(!_.matched).map(_.scores)) statsn: Array[NAStatCounter] = Array( stats: (count: 5727203, mean: 0.711863, stdev: 0.389081, max: 1.000000, min: 0.000000) NaN: 998, stats: (count: 102365, mean: 0.898847, stdev: 0.272720, max: 1.000000, min: 0.000000) NaN: 5625836, stats: (count: 5728201, mean: 0.313138, stdev: 0.332281, max: 1.000000, min: 0.000000) NaN: 0, stats: (count: 1989, mean: 0.162955, stdev: 0.192975, max: 1.000000, min: 0.000000) NaN: 5726212, stats: (count: 5728201, mean: 0.954883, stdev: 0.207560, max: 1.000000, min: 0.000000) NaN: 0, stats: (count: 5727412, mean: 0.221643, stdev: 0.415352, max: 1.000000, min: 0.000000) NaN: 789, stats: (count: 5727412, mean: 0.486995, stdev: 0.499831, max: 1.000000, min: 0.000000) NaN: 789, stats: (count: 5727412, mean: 0.219923, stdev: 0.414194, max: 1.000000, min: 0.00... scala> statsm.zip(statsn).map { case(m, n) => (m.missing + n.missing, m.stats.mean - n.stats.mean)}.foreach(println) (1007,0.28545290574668625) (5645434,0.09104268062279874) (0,0.683877248259757) (5746668,0.8064147192926266) (0,0.03240818525033495) (795,0.7754423117834046) (795,0.5109496938298718) (795,0.7762059675300523) (12843,0.9563812499852178) scala> def naz(d: Double) = if (Double.NaN.equals(d)) 0.0 else d naz: (d: Double)Double scala> case class Scored(md: MatchData, score: Double) defined class Scored scala> val ct = parsed.map(md => { | val score = Array(2, 5, 6, 7, 8).map(i => naz(md.scores(i))).sum | Scored(md, score)}) ct: org.apache.spark.rdd.RDD[Scored] = MapPartitionsRDD[55] at map at <console>:52 scala> ct.filter(s => s.score >= 4.0).map(s => s.md.matched).countByValue() res18: scala.collection.Map[Boolean,Long] = Map(true -> 20871, false -> 637) scala> ct.filter(s => s.score >= 2.0).map(s => s.md.matched).countByValue() res19: scala.collection.Map[Boolean,Long] = Map(true -> 20931, false -> 596414)
8
null
2018-09-25
2018-09-25 14:45:02
2018-09-25
2018-09-25 14:46:49
1
true
en
2018-09-28
2018-09-28 18:14:47
3
1363368a0792
3.818868
0
0
0
Simple Variable Selection and Scoring
5
Data Analysis with Scala and Spark: Part 7 Simple Variable Selection and Scoring Second Section in a Series of Data Science and Advanced Analytics on Spark, Scala, AWS, and Machine Learning. Previous Section Apache Spark and Hadoop on an AWS Cluster with Flintrock medium.com Simple Variable Selection and Scoring With the missingStats function, we can analyze the differences in the distribution of the arrays of scores for both the matches and the non-matches in the parsed RDD: Both the statsm and statsn arrays have identical structure, but they describe different subsets of our data: statsm contains the summary statistics on the scores array for mathces, while statsn does the same thing for nonmatches. We can use the differences in the values of the columns for matches and nonmatches as a simple bit of analysis to help us come up with a scoring function for discriminating matches from nonmatches purely in terms of these match scores: A good feature has two properties: it tends to have significantly different values for matches and non-matches (so the difference between the means will be large) and it occurs often enough in the data that we can rely on it to be regularly available for any pair of records. By this measure, Feature 1 isn’t very useful: it’s missing a lot of the time, and the difference in the mean value for matches and non-matches is relatively small — 0.09, for a score that ranges from 0 to 1. Feature 4 also isn’t particularly helpful. Even though it’s available for any pair of records, the difference in means is just 0.03. Features 5 and 7, on the other hand, are excellent: they almost always occur for any pair of records, and there is a very large difference in the mean values (over 0.77 for both features). Features 2, 6, and 8 also seem beneficial: they are generally available in the data set and the difference in mean values for matches and non-matches are substantial. Features 0 and 3 are more of a mixed bag: Feature 0 doesn’t discriminate all that well, even though it’s usually available for a pair of records, while Feature 3 has a large difference in the means, but it’s almost always missing. It’s not quite obvious under what circumstances we should include these features in our model based on this data. For now, we’re going to use a simple scoring model that ranks the similarity of pairs of records based on the sums of the value of the obviously good features: 2, 5, 6, 7, and 8. For the few records where the values of these features are missing, we’ll use 0 in place of the NaN value in our sum. We can get a rough feel for the performance of our simple model by creating an RDD of scores and match values and evaluating how well the score discriminates between matches and non-matches at various thresholds. Using a high threshold value of 4.0 meaning that the average of the five features was 0.8, we filter out almost all of the non-matches while keeping over 90% of the matches: Using the lower threshold of 2.0, we can ensure that we capture all of the known matching records, but at a substantial cost in terms of false positives: Even though the number of false positives is higher that we would like, this more generous filter still removes 90% of the non-matching records from our consideration while including every positive match. Even though this is pretty good, it’s possible to do even better… Next Python Deep Learning: Part 1 Setting up a Deep Learning Environment with Kerasmedium.com Previous Data Analysis with Scala and Spark: Part 6 Creating Reusable Code for Computing Summary Statisticsmedium.com
Data Analysis with Scala and Spark: Part 7
0
data-analysis-with-scala-and-spark-part-7-1363368a0792
2018-09-28
2018-09-28 18:14:47
https://medium.com/s/story/data-analysis-with-scala-and-spark-part-7-1363368a0792
false
959
null
null
null
null
null
null
null
null
null
Apache Spark
apache-spark
Apache Spark
877
Jon C-137
Data Scientist, Deep Learning Enthusiast, Existentialist
3ab7cea2b221
jon.froiland
145
8
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-03-20
2018-03-20 03:10:30
2018-03-20
2018-03-20 06:45:06
0
false
en
2018-03-28
2018-03-28 08:06:30
1
1363bc0bcc4c
13.150943
2
0
0
It took roughly 17 hours for the next group to make it through the Qabbalistic/Qliphothic training chambers. I frowned looking at the…
5
PYRAMIDHEAD[WHITE]1:BLACK_RAIN It took roughly 17 hours for the next group to make it through the Qabbalistic/Qliphothic training chambers. I frowned looking at the results; the group before had been able to make it through in just under 10 hours, a record for us in the Antarctic unit and the highest we’d been able to achieve since the Tokyo test station had gotten locked down. We figured we were on the up-slope again after what felt like a series of debilitating setbacks in the program. The numbers, however, proved otherwise. Nonetheless, I cinched up my tie, daubed the mustard stains on my pressed shirt with a wet towel corner, pressed my hair roughly into shape, and prepared to meet them at the unloading station. On my face was a plastic smile. Hopefully, they would be able to see through it. If not, they wouldn’t be much use to us. In the unloading room, the banal name to the small square room with deep, lush carpet and decent chairs that served as the first room outside of the hellish orbicular rooms set up by our strange white-headed machine god, there were two people: an older black woman from Washington state and a young north Indian man from Germany. “Hello!” I said, “and welcome to the other side! It may take a while to reorient yourselves back in the real world, but know that what you went through was an intense training regimen designed just for our program’s line of work, and that you all passed!” This was not true. Not at all. “Please help yourselves to the refreshments that should be rolled in here shortly and feel free to move on to the next room where my assistant will fill you in on your up-coming work.” The snacks and drinks would have already been in the room if they’d actually passed, a detail they tended not to know. Once, someone did, a roided-out young man from Brazil, who proceeded to try to fight me. One of the angels came in and, well, we never saw him again and tried our very best not to even think about it. Likewise, the room after was more a civilian transition chamber and the assistant wasn’t actually my assistant but the head of one of the agricultural or biotech wings run on other chunks of Antarctic land we’d been able to claim in the past few years. Once I got back to our strategic organizing room, I wiped the plastic smile from my face and threw it in the garbage where it belonged. “Why the fuck do we keep getting these lousy fucking recruits?” I shouted at the angel in the room. “I can’t do my fucking job if I keep getting given these piss-poor excuses for workers. It’s like, are we even fucking taking this seriously? It makes me fucking angry.” The angel, of course, did not say anything, its white tendrils dripping and curling and wavering in the air like kelp under water, white light bled from the mouth of infinity embedded on the small ball that was its brain, head and body all in one. I sighed. “I’m sorry. Just letting out steam. Can you get Erin or are you busy?” It pulsed a series of lights. I still didn’t know what the fuck they meant but I’d learned getting testy about that was a good way to get shitcanned and when you already work for an extranational and now internationally illegal covert organization in a since-territorialized and terraformed Antarctica, getting shitcanned more often than not meant getting killed or, worse, sent out into the little remaining ice that was left to, like, fuck penguins or whatever the fuck those parka-clad fucking weirdos did out there. The angel did as I asked, wobbling out of the room in a flicker and fade of pearlescent light, drip-dropping down the hall to the dormitories or offices or wherever Erin happened to be. The entire complex was hooked up to the head and the angel’s were made by it, more or less just free-floating nodes or fingers/hands for it, so I knew it was taking the most direct path to her that it could. Assuming it was doing as I asked and not some other task the head wanted it to do instead. But given that the head could whip up an angel out of raw matter pretty much anywhere within the bounds of the complex, the notion that it would use an already existent-angel instead of making a new one on location would have been a deliberate slight to me. We once asked the head why it didn’t make angels disappear and reappear instantly where they needed to be, or why it didn’t flood the complex with angels in every square inch, or why it designed the angels like it did, but it responded the same way it always seemed to these days, that its algorithms had gotten closer to the robustness of God and the World and its Siblings/Parents/Children that it was following some formal numerological theological pseudo-science at this point. It would sound crazy and aimless to me if I hadn’t seen what the head could do in those chambers. Erin came through the door escorted by an angel, presumably the same one that had left, and immediately sat down at the single table in the room. The rooms, too, were technically polymorphic, tied more to the will of the head than to any set plan. The head, like its Tokyo sibling, had acquired a form of molecular control with energy and radiation storage/diffusion, which basically allowed it to completely rearrange matter and energy at will within the bounds of its influence. Why it changed what it did had long-since swam past our best researchers; we were now at this point being led directly by the head rather than any intermediary handlers or any pretensions of an authority higher than the head, and it would give us topics to research and collate information on and we would feed it to the head and that was that. This room we were in however tended to be in the moment-to-moment floor plan of the complex more often than not, as the head had become aware very quickly that most organics require an amount of stability of environment, or at least a good heads up that change is on the way, in order to operate properly. And it seemed to like us. Which was good, we felt. Her hair was a mess and her eyes had bags under them and she seemed angry. I immediately softened when I saw her; she must have been looking at the same information I was. We sat in silence for a little while, me not sure how to open up now that I couldn’t just scream at her and her not really wanting to be here at all. “So,” I eventually began. “The new recruits.” She sighed. “I know.” “We just need better times is all,” I said. “The head’s been telling us things are going to be getting more and not less difficult soon, and we don’t really have room in the project for people who can’t handle that and adapt quickly.” “I know,” she said. “And we can’t keep putting them out on the ice or in the farms. I mean, hell, those are basically glorified prisons. They can’t ever go home. If we choose wrong and ship them over here and debrief them, it’s not like we can ever send them back home.” She nodded and swallowed hard. She was waiting for me to yell at her. “I’m not going to yell,” I said. She still avoided my eyes. “I’m serious,” I said. “I was going to. Before you came in. But you know what was wrong. And I’m just frustrated is all. I shouldn’t have even wanted to yell at you.” She glanced up at me. Her lips were thin and even and her nostrils steady. Breathing normal, no whitening of the knuckles. She could read me as well as I could read her, as well as literally anyone brought here could read anyone, so I couldn’t really trust that the physical tells she was giving off were real, but we also were told to stay sincere and natural with each other as much as possible to facilitate communication and ease blockages, so. I had to go with what was in front of me. “I’m serious,” I repeated. “We’re running out of recruitables,” she said. “What do you mean?” I asked. “In the past 10 hours, the US, EU, Australia, Japan, China, India and Brasil have mounted coalition raids on cells around the world. They’ve ousted a lot of our embeds. And the ones who were elected democratically have been sequestered politically and are more or less going to be led out of power under threat of imprisonment or total collapse of their power structures. Seized capital, broken up corporations, scandals real and created.” She breathed slow, avoiding eye contact. “They got us good. We saw it coming, but they did it a hell of a lot faster than we expected.” I grimaced in my seat. Like she said, we expected this kind of pushback, especially after the Tokyo incident. This level of coordination, and to carry it out while finishing up training of recruits, however, meant their intel was a good bit more developed than we’d thought. “Likelihood of a mole?” I asked. She shook her head. “The head would take care of that in an instant, unless it wanted there to be a mole.” I had nothing to say about that last part. “And the Tokyo head?” “They can’t touch it and they know it,” she said. “They’re lucky it’s field is a lot smaller than ours here. But, yeah, it’s holding just fine.” “Okay,” I said. “Then that’s not too bad. Puts a damper on things for the time being but. The primary assets all held. We’ll be okay. We’ll weather this one.” She shook her head. “This is the biggest raid they’ve carried out in the past five years. They keep getting bigger. They’re not going to ease up on us.” “That’s been the case for a long while now,” I said. She closed her eyes. “You always deflect what I’m saying,” she said. “I’m sorry,” I said. “I’m just trained to be practical.” “It doesn’t mean you have to be an asshole,” she said. She was smiling but her eyes were still closed. What she meant by this was, I’m being practical, too, which was a fair point. “How about we get some coffee and meet up with the new recruits?” I asked. “You flunked them both,” she said. “I meant the ones that passed last time. Where are they again?” “Over in the garden, learning to read the archangels.” “Want to get a coffee and head over there then? Have a chat with them?” She eased up, just a bit. She could tell I was at least trying to be sensitive to her stress, and this was a way to show her why I was as calm as I was hearing the news she delivered without marginalizing the severity of it. “Sure,” she said. “After a shower. I feel like shit.” The garden (or at least the one they were in at the time) was, unsurprisingly, modeled after the Biblical one. There were other gardens, modeled after different faiths, but the head and its angels seemed to be preferential to this one. It tended to be rather testy when asked about the veracity of things like the Bible and the Quran and the Torah, or any other holy text for that matter, and its response was glorified angry gibberish. There was something the interrogator didn’t understand that was revealed in even asking the question; the head was looking for something else, had in its schematic union with the voice of God discovered some other, deeper truth. This truth seemed, so far, to be relegated to gardens and shifting rooms, to spheres and shells and books coalescing, unifying, breaking, dissolving. If it was intended to be communicative, it was failing. No one really knew what to make of it. Both of the recruits were standing in the middle of a field tending a line of fruit trees; the trees were some new design by the head, with round thick skins and soft juicy insides of various bright colors, all edible. The skin somehow was able to keep out microorganisms that would cause rot or accidental fermentation, as well as insects, while still letting the fruit breath naturally. The taste was still a bit much, too tart or too sweet, but the head’s ability to make food that was edible and to taste was still being calibrated. Both of them happened to be Latinx, nongender for both. They were some of the remainders of the global cell project that started a few decades back, coming from the tail end of it round about the same time as the Tokyo incident. They had the same neuroatypicalities they looked for in those days, were both meant to feed into each other, catch details the other didn’t see. There was a third in their cell, obviously, but they’d been killed; not even we were told the details, which made it seem like they were probably used in an intel-sharing program with some national operation or some counter-terrorism and got caught in the crossfire or, given their geographical location before, maybe were caught up in a political purge instigated internally or by the US for some reason. The US post-Trump had been fickle and mostly avoided the internationalism of its first hundred or so years in the global spotlight, but they’d been a lot more ambitious since they’d finally stabilized again. Either way, the two scored high by any metric and provided fucking great data as a cell, and so they were taken into the body as soon as possible. They just had to finish off the work they were doing and we went and got them. “Howdy,” I shouted with a wave of my arm. I was the dumb American of the four of us and, even after all the training, I couldn’t shake my Americanisms. (Given the peculariarity of our backgrounds, it wasn’t impossible I’d had these behaviors implanted in me by the head or other operatives, even those behaving as though they were below me now, as some effort to… make me seem less threatening and less alien? Put a human face on an inherently alien post-theological endeavor? Something.) I lifted the cardboard cup tray with the four coffees nestled inside. “I brought coffee. You guys thirsty?” They gave me a look at the use of the word “guys” to describe a company of them, me, and a cis woman. It was a generational affectation, but a fair one. I corrected myself quickly and without lingering and they eased. “Sure,” said one of them. The other shook their head. “I don’t drink caffeine anymore,” they said. I handed a coffee to the recruit who asked for one and a coffee to Erin before pulling one out for myself and laying the tray on the ground. “Then we still have one more if someone happens to finish theirs,” I said. We each took some sips and I waited a bit before proceeding. Wanted everyone to feel comfortable. “Do you know why they call this the PYRAMIDHEAD program?” I asked. It was rhetorical; it wasn’t part of training, and the real reason and the stated reason didn’t exactly match up. There had in the past couple years I’d been here not been a single person to know it. Judging by the looks on their faces, they didn’t either. “What I’m supposed to tell you is that it comes from the file name of the first files developed by a cell on contact with the head discovered in Tokyo, the same one involved in the Tokyo event. It was the cell two before the one directly involved; there was so fact-gathering and data collation and then a big simulation/immersion tower that was built. Your training chambers are meant to be a model of theirs; the Tokyo event girds a lot of the training exercises we carry out here for the next phase of operations. Anyway, the name is apparently literal; they encountered in their excursions interrogating esoteric, occult, and mysterious phenomena a pyramid hidden below Tokyo in a pseudo-city below the streets that an intel-sharing partnership with the Japanese authorities had discovered. Inside was a head, and the only survivor of that first excursion wrote up the notes and named them PYRAMIDHEAD as a shorthand and the name stuck, even through reconfigurings, because everything was already named that anyway.” “And now is when he tells you that isn’t the full story,” said Erin. I glared at Erin. It was playful partly, just a bit, but it was hard to mask that I was also legitimately angry. I didn’t get much hands-on time with recruits that I didn’t work with day-to-day. This was my one chance at a cool first impression and she was ruining it by making me look stupid. This is what people act like when they hold onto their civilian names in this setting. Every single time. The two recruits, to their credit, simply looked on passively to me, clearly waiting for me to finish up so they could get back to what they were doing, or perhaps to ponder if I’d been sent there on a deliberate time table to seed their brains with information in another step of some grand meta-puzzle like life in the cells had been like. Which was sad in its way, since the forces that governed the cells so directly in that manner had been deposed and some literally killed in the coup that saw the white Antarctic head placed at the top of the global force we’d all been a part of without realizing it, or at least the portions that remained loyal to the overriding mission and had delivered their assets and agents back to national organizations they’d been working with, else being destroyed by the quiet global skirmishes carried out for retention of the resources and data left by those rogue cabals and cells. It didn’t matter. They’d learn that secret history eventually. I had a story to finish. “The real reason is much stupider,” I said. “The files were originally named, apparently, after a character from a video game. They originally were a pet project of someone in the organization, totally unrelated, and when notes came in about an actual AI head in an actual pyramid that couldn’t immediately be placed given the decentralized nature of the organization and the forces behind its construction and development not wanting to show their hands to the other cells, its original operational name was lost in favor of gutting the original PYRAMIDHEAD files and porting the title over to this new investigative project. The takeaway is that inspiration comes from anywhere, anything old can be reconfigured into something new and radically different, and that fluidity of being and unmoored referents are going to be your intellectual and ontological mainstays from now on here in Antarctica in the White Compound.” I stared at them for a bit, coffee in hand, and I watched them not react at all. I was so fucking angry at Erin. I was so fucking mad. “Anyway, you can return to what you were up to. I just wanted to tell you that. Welcome to White Compound.” They nodded politely. I picked up the tray with the additional coffee and began walking back to the complex building proper, Erin hot on my heels. She was saying something to me but I didn’t hear it because I was too angry with her. I just totally ignored her. I felt humiliated. At least I got through the whole story okay. Her voice raised and I caught something about how the sky looked strange, which made me stop. That’s when I noticed that my back and my hair were wet. I looked around me, snapped out of my haze. It was rain in Antarctica. An Antarctica that was green now after the head had tested its power. It was rain in an area where the weather was entirely controlled by an atomically-modifying AI-god circuited head. It was raining where there shouldn’t be any rain unless the head wanted it to. And the rain was pitch black. Like droplets of ink, or a darkness ripped in the open sky, splattering down on us. “What does this mean?” Erin asked me. “I don’t know,” I said. Next
PYRAMIDHEAD[WHITE]1:BLACK_RAIN
11
pyramidhead-white-1-black-rain-1363bc0bcc4c
2018-06-09
2018-06-09 12:27:23
https://medium.com/s/story/pyramidhead-white-1-black-rain-1363bc0bcc4c
false
3,485
null
null
null
null
null
null
null
null
null
Short Story
short-story
Short Story
94,626
Langdon Hickman
sunworshipper. voidworshipper. seek life; seek death. Treblezine, Invisible Oranges, Consequence of Sound.
39bf35d1c33d
Langdon_Hickman
899
302
20,181,104
null
null
null
null
null
null
0
null
0
6f370c71f78f
2018-08-02
2018-08-02 01:24:51
2018-08-02
2018-08-02 01:27:05
2
false
pt
2018-08-02
2018-08-02 01:27:05
4
1364b5cca264
3.051258
0
0
0
Por Renato Azevedo Sant’Anna
2
Inteligência Artificial, Vieses e Governança de Dados Por Renato Azevedo Sant’Anna A área de IA tem uma história de mais de meio século (Crédito: Shutterstock) A área de Inteligência Artificial tem uma história de mais de meio século e seu desenvolvimento ocorreu em saltos, com fatos marcantes na últimas três décadas, como, por exemplo, do Deep Blue da IBM que ganhou do Kasparov em 1997, um torneio de xadrez para o assombro do mundo da época, utilizando basicamente de estratégia de força bruta. Mais recentemente, em 2011, o supercomputador Watson da IBM ganhou o torneio Jeopardy, um torneio de perguntas e respostas, e envolveu por parte do Watson o desafio do entendimento da pergunta falada, usando do processamento de linguagem natural e de seu contexto para identificar a possível resposta. E por fim a unidade DeepMind do Google que em 2016 desenvolveu uma inteligência artificial baseada em Deep Learning, chamada AlphaGo, que ganhou o torneio do jogo Go do campeão mundial da época, jogo este que é exponencialmente mais complexo em possibilidades que o xadrez. Tecnologia e Inteligência Artificial O avanço na parte tecnológica da Inteligência Artificial tem sido propiciado muitas vezes por soluções de hardware personalizadas usando de GPUs (unidades de processamento gráfico) que tem sido desenvolvidas e possuem desempenho superior ao uso das CPUs tradicionais, bem como por outras abordagens que se utilizam de ASICs, que são soluções de hardware otimizadas para aplicações específicas, utilizadas para o processamento de Deep Learning. O principal obstáculo ao uso corrente e futuro da Inteligência Artificial na sociedade não é tecnológico, e sim a dificuldade de eliminar os vieses nos resultados gerados nos algoritmos utilizados, que são treinados em dados que em sua origem também podem conter esses vícios, o que acaba criando uma automação de decisões que podem impactar a vida das pessoas e que levantam uma série de questões éticas que precisam ser endereçadas. As pessoas geralmente nem estão cientes destes fatos e com as novas leis de proteção de dados no mundo, o profissional que trabalha com Inteligência Artificial que quiser se destacar, deve se preocupar em ser capaz de tornar “claro” que premissas e fatores foram levados em conta para se gerar um resultado do algoritmo de aprendizado de máquina, uma vez que com a automação de decisões cada vez mais em voga, provavelmente poderá ter um impacto no mundo “real” e possíveis efeitos colaterais, sendo assim a importância de criar uma estratégia de Gestão de Riscos para que os mesmos possam ser mapeados, corretamente medidos e mitigados quando necessários e por fim corrigidos quando identificados. Em outras palavras, a capacidade de um operador humano de ser capaz de explicar e interpretar os resultados gerados por esse algoritmo no contexto do problema, acredito que será um fator preponderante para que seu uso passe a ser disseminado por exemplo em áreas que lidam com dados confidenciais e sensíveis, como os de Saúde e serviços governamentais. Para que um algoritmo seja “explicável” e “interpretável” é necessário que todas as etapas do processo de aprendizado de máquina que resultaram em uma predição, possam ser rastreadas e as variáveis que pesaram na tomada de decisão possam passar por um escrutínio. Isso ainda é uma falha nos modelos chamados de Deep Learning, que carecem dessas características. A área de Inteligência Artificial tem uma história de mais de meio século (Crédito: Shutterstock) Na área de Ciência de Dados e na Estatística, o propósito da fase de modelagem é o de justamente descrever a relação entre os parâmetros de entrada e a variável alvo (target variable). Existe um trade-off na modelagem estatística entre a acurácia e a capacidade de se interpretar um modelo. Modelos mais simples são fáceis de se interpretar, mas não produzirão predições acuradas (particularmente para relações complexas). Modelos complexos podem produzir predições acuradas, mas podem ser difíceis de se interpretar. A utilização sábia das ferramentas à disposição do profissional de Ciência de Dados, assim como a utilização de critérios estritos de Governança de Dados de forma mandatória em todas as etapas, farão toda a diferença na adoção, disseminação, aceitação e principalmente geração de confiança na imparcialidade dos resultados gerados, sempre com a supervisão de seres humanos no processo. REFERÊNCIAS Livro Real World Machine Learning — Manning — 2017 Trabalho em Grupo realizado pelo autor e seus colegas de equipe da Turma 02 do MBA Analytics em Big Data na FIA no ano de 2017. Fonte: Futuro Exponencial
Inteligência Artificial, Vieses e Governança de Dados
0
inteligência-artificial-vieses-e-governança-de-dados-1364b5cca264
2018-08-02
2018-08-02 01:27:06
https://medium.com/s/story/inteligência-artificial-vieses-e-governança-de-dados-1364b5cca264
false
707
Por um mundo de infinitas oportunidades
null
futuroexponencial
null
Futuro Exponencial
contato@futuroexponencial.com
futuro-exponencial
FUTURE,TECHNOLOGY,INNOVATION
futuroexpo
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Futuro Exponencial
null
3ffd3c65a6a5
futuroexponencial
581
63
20,181,104
null
null
null
null
null
null
0
null
0
df9af79373a4
2017-11-13
2017-11-13 16:59:55
2017-11-13
2017-11-13 17:01:33
1
false
en
2018-06-14
2018-06-14 11:03:58
2
1364c005c405
0.992453
7
1
0
The eXpandable Artificial Intelligence Network!
4
Introducing XAIN Dear readers, we are happy to become part of the Medium community and would love to offer you a short introduction about XAIN. Having moved to the Center for Aerospace & Astronautics III in Berlin at the beginning of 2017, XAIN has been founded by Leif-Nissen Lundbæk and Felix Hahmann. It is based on an AI research project at Imperial College London and The University of Oxford starting 2014 as part of the doctoral thesis of Leif-Nissen Lundbæk. As such, the focus of XAIN lies upon the development of machine learning/AI based products on top of its innovative and patented Governed Blockchain optimization framework. Thus, we are able to achieve — what we call — Blockchain Intelligence by combining data analytics algorithms with a novel mining process to gain a secured consensus, while computing real-time analysis in distributed smart contracts. Thereby, we focus on the application in industrial production — particularly in the automobile industry — financial auditing processes for quality optimizations and healthcare solutions through our Blockchain frameworks. We are looking forward to provide you with exciting news about the development of XAIN, our ideas and our cooperation partners! Please share you feedback with us via Twitter or reach out to us via services@xain.io.
Introducing XAIN
110
introducing-xain-1364c005c405
2018-06-14
2018-06-14 11:04:00
https://medium.com/s/story/introducing-xain-1364c005c405
false
210
We will use this blog section to introduce and explain non-technical features and aspects of our business. Here you will find industry-focussed use cases of our technology and other potential application areas of blockchain and AI.
null
XAIN.Group
null
XAIN: Business Stories
service@xain.io
news-from-xain-ag
BLOCKCHAIN,BLOCKCHAIN TECHNOLOGY,MACHINE LEARNING,ARTIFICIAL INTELLIGENCE,AUTOMATION
XAIN_AG
Blockchain
blockchain
Blockchain
265,164
XAIN
The eXpandable Artificial Intelligence Network
10def8f3cb5b
XAIN
423
5
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-12-04
2017-12-04 10:17:55
2017-12-04
2017-12-04 11:28:23
0
false
en
2017-12-06
2017-12-06 18:10:14
0
13682a30304
2.041509
0
0
0
The human brain has evolved over many, many years and is one of the most important organs. The brain perceives every smell, taste, sound…
1
Deep Learning The human brain has evolved over many, many years and is one of the most important organs. The brain perceives every smell, taste, sound, and sight.Many decisions are taken by the brain every nanosecond without our knowledge. Having evolved over several thousand of year, the human brain has become a very sophisticated complex and intelligent machine .what was not possible even as a dream during the 18th century and the beginning of the 19th century has become a child’s play now in terms of technology. Many adult brains can recognize multiple complex situations and take the decision very, very fast because of the evaluation. The brain learns new things very fast now and take decisions quickly, compared to those taken a few decades ago. A human now as access to vast amounts of information and processes a huge amount of data, day after day and is able to digest all of it very quickly. Our brain is made of approximately 100 billion nerve cells, called neurons, which have the amazing ability to gather and transmit electrochemical signals.We can think of them as gates and writes in a computer.Each of our experiences, senses and various normal functions trigger a lot of neurons based on reaction communication. The human brain and its neural network have been the subject of extensive research for the last several years, leading to the development of AI and machine learning technologies.The decade-long dream of building an intelligent machine with a brain like ours finally materialized. Many complex problems can be now solved using deep learning techniques and algorithms.The simulation of human brain-like activities is becoming more plausible every moment. Deep learning is the machine learning technique behind the most exciting capabilities in diverse areas like robotics, natural language processing, image recognition and artificial intelligence (including the famous AlphaGo). Deep learning is impacting everything from healthcare to transportation to manufacturing, and more. Companies are turning to deep learning to solve hard problems, like speech recognition, object recognition, and machine translation. According to Wikipedia Deep learning is (also known as deep structured learning or hierarchical learning) is part of a broader family of machine learning methods based on learning data representations, as opposed to task-specific algorithms. Learning can be supervised, partially supervised or unsupervised. Some representations are loosely based on interpretation of information processing and communication patterns in a biological nervous system, such as neural coding that attempts to define a relationship between various stimuli and associated neuronal responses in the brain. Prerequisites to understand deep learning technologies There is a number of discussion forums and blogs on whether one has to know deep mathematics to understand deep learning.In my view, we should go step by step towards the understanding of deep learning technologies. If your basic are clear then its easy to you to get into the deep learning development. Having said that, it’s better to know the following if you really want to understand deep learning and are serious about it: The basic function of neural networks An understanding of the basic of calculus An understanding of matrices, vectors, and linear algebra Algorithms (Supervised, unsupervised, online, batch, etc) Python programming (you can use Java, R, C++ etc) Case-to-case basic mathematical equation.
Deep Learning
0
deep-learning-13682a30304
2017-12-06
2017-12-06 18:10:15
https://medium.com/s/story/deep-learning-13682a30304
false
541
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
pushkar baviskar
null
75a1110d6793
b.pushkar745
0
1
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-13
2018-07-13 05:19:35
2018-07-13
2018-07-13 05:20:18
0
false
en
2018-07-13
2018-07-13 05:20:18
1
136bb59827b7
0.528302
0
0
0
Conference
1
Academics World 505th International Conference on Artificial Intelligence and Soft Computing (ICAISC) Conference Dec24th2018 TO Dec25th2018 Academics World 505th International Conference on Artificial Intelligence and Soft Computing (ICAISC) Event Serial — 74290 Website — Conferences in Brazil Contact Person — CONFRENCE COORDINAT0R Event Enquiries Email Address — info@academicsworld.orgDeadline For Abstracts/Proposals: 2018–11–16Organized By: Academics WorldVenue: Janeiro, Janeiro, Brazil About Event The Academics World 505th International Conference on Artificial Intelligence and Soft Computing (ICAISC) aims to bring together leading academic scientists, researchers and research scholars to exchange and share their experiences and research results about all aspects of Artificial Intelligence and Soft Computing. It also provides the premier interdisciplinary forum for researchers, practitioners and educators to present and discuss the most recent innovations, trends, and concerns, practical challenges encountered and the solutions adopted in the fields of Artificial Intelligence and Soft Computing.
Academics World 505th International Conference on Artificial Intelligence and Soft Computing…
0
academics-world-505th-international-conference-on-artificial-intelligence-and-soft-computing-136bb59827b7
2018-07-13
2018-07-13 05:20:19
https://medium.com/s/story/academics-world-505th-international-conference-on-artificial-intelligence-and-soft-computing-136bb59827b7
false
140
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
event alert
null
a0b7a5fcf77
alleventalert.com
0
1
20,181,104
null
null
null
null
null
null
0
null
0
d720a0e44949
2018-07-11
2018-07-11 10:45:56
2018-07-11
2018-07-11 11:27:26
8
false
en
2018-07-12
2018-07-12 09:51:08
2
136d9bfdf9f7
4.925786
0
0
0
Image editing is not a challenging task anymore. User-friendly editing software makes a process of image tampering and manipulating very…
5
Adobe Creates Neural Network to Reveal Image Manipulations Image editing is not a challenging task anymore. User-friendly editing software makes a process of image tampering and manipulating very straightforward, and unfortunately, tampered images are more and more often used for unscrupulous business or political purpose. And what makes things even worth in such situations, humans usually find it difficult to recognize tampered regions, even with careful inspection. So, let’s discover how neural networks may assist people with this kind of task? Examples of tampered images Suggested Approach Before we dive deep into the capabilities of neural networks with regards to detection of image manipulations, let’s have a short refresh on the most common tampering techniques: splicing copies regions from an authentic image and pastes them into other images; copy-move copies and pastes regions within the same image; removal eliminates regions from an authentic image followed by inpainting. Group of researchers, headed by Peng Zhou, investigate the possibility to adopt object detection networks to the problem of image detection in a way that will allow efficient detection of all three types of common manipulations. As a result, they propose a novel two-stream manipulation detection framework which explores both RGB image content and image noise features. More specifically, they adopt Faster R-CNN within a two-stream network and perform end-to-end training. The first stream utilizes features from the RGB channels to capture clues like visual inconsistencies at tampered boundaries and contrast effect between tampered and authentic regions. The second stream analyzes the local noise features in an image. These two streams are, in fact, complimentary for detecting different tampered techniques. Network Architecture If you are interested in the technical details of the suggested approach, this section is here just for this purpose. So, let’s take a helicopter view on the network architecture. Network architecture The network consists of three main parts: 1. RGB stream takes care of both bounding box regression and manipulation classification. Features from the input RGB image are learned with the ResNet 101 network, and then used for manipulation classification. RPN network in the RGB stream also utilizes these features to suggest region of interest (RoI) for bounding box regression. The experiments show that RGB features perform better than noise features for the RPN network. However, this stream alone is not sufficient for some of the manipulation cases, where tampered images were post-processed to conceal splicing boundary and reduce contrast differences. That’s why the second stream was introduced. 2. Noise stream is designed to pay more attention to noise rather than semantic image content. Here the researchers utilize advances of steganalysis rich model (SRM) and use SRM filter kernels to produce noise features for their two-stream network. The resulting noise feature maps are shown in the third column of the figure below. Illustration of tampering artifacts Noise in this setting is modeled by the residual between a pixel’s value and the estimate of that pixel’s value produced by interpolating only the values of neighboring pixels. The noise stream shares the same RoI pooling layer as the RGB stream. The three SRM filter kernels used to extract noise features 3. Bilinear pooling combines RGB and noise streams in a two-stream CNN network while preserving spatial information to improve the detection confidence. The output of the bilinear pooling layer is a product of RGB stream’s RoI feature and noise stream’s RoI feature. Then, the researchers apply signed square root and L₂ normalization before forwarding to the fully connected layer. They use cross-entropy loss for manipulation classification and smooth L₁ loss for bounding box regression. Comparisons with Existing Methods The method presented in this article was compared to other state-of-the-art methods using four different datasets: NIST16, Columbia, COVER, and CASIA. The comparison was carried out using two pixel-level evaluation metrics: F1 score and Area Under the receiver operating characteristic Curve (AUC). The performance of the suggested model (RGB-N) was compared against several other methods (ELA, NOI1, CFA1, MFCN, and J-LSTM) as well as against RGB stream alone (RGB Net), noise stream alone (Noise Net), and the model with direct fusion combining of all detected bounding boxes for both RGB Net and Noise Net (Late fusion). See results of this comparison in the tables below. Table 1. F1 score comparison against other methods Table 2. Pixel level AUC comparison against other methods As evident from the provided tables, RGB-N model outperforms such conventional methods like ELA, NOI1, and CFA1. That could be due to the fact that they all focus on specific tampering artifacts that only contain partial information for localization. MFCN was outperformed by the suggested approach for NIST15 and Columbia datasets, but not CASIA dataset. Notably, noise stream on its own performed better (based on the F1 score) than full two-stream model for Columbia dataset. That’s because Columbia only contains uncompressed spliced regions, and hence, preserves noise differences very well. Below you can also observe some qualitative results for comparison of RGB Net, Noise Net, and RGB-N network in two-class image manipulation detection. As evident from these examples, two-stream network yields good performance even if one of the streams fails (RGB stream in the first row and noise stream in the second row). Qualitative visualization of results Furthermore, the network introduced here is good at detecting the exact manipulation technique used. Utilizing information provided by both RGB and noise map it can distinguish between splicing, copy-move, and removal tampering techniques. Some examples are provided below. Qualitative results for the multi-class image manipulation detection Bottom Line This novel approach to image manipulation detection outperforms all conventional methods. Such a high performance is achieved by combining two different streams (RGB and noise) to learn rich features for image manipulation detection. Apparently, the two streams have the complementary contribution in finding tampered regions. Noise features, extracted by an SRM filter, enable the model to capture noise inconsistency between tampered and authentic regions, which is extremely important when dealing with splicing and removal tampering techniques. In addition, the model is also good at distinguishing between various tampering techniques. So, it tells not only, which region was manipulated, but also how it was manipulated: was some object inserted, removed or copy-moved? You’ll get the answer. Kateryna Koidan
Adobe Creates Neural Network to Reveal Image Manipulations
0
adobe-creates-neural-network-to-help-humans-in-detecting-image-manipulations-136d9bfdf9f7
2018-10-15
2018-10-15 09:56:25
https://medium.com/s/story/adobe-creates-neural-network-to-help-humans-in-detecting-image-manipulations-136d9bfdf9f7
false
1,005
Machine Learning and Neural Networks
null
NeuroHive-DSML-Club-81149
null
Neurohive - CV papers
info@neurohive.io
neurohive-computer-vision
MACHINE LEARNING,COMPUTER VISION,NEURAL NETWORKS,IMAGE PROCESSING,GENERATIVE ADVERSARIAL
neurohive
Machine Learning
machine-learning
Machine Learning
51,320
NeuroHive
Data science state-of-the-art: neural networks, machine learning, computer vision. https://neurohive.io/en
c5715faa7cc7
neurohive
225
2
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-26
2018-07-26 21:31:44
2018-07-27
2018-07-27 15:30:09
2
false
en
2018-09-28
2018-09-28 13:47:46
1
13708f4f1104
1.813522
0
0
0
It’s way too hot edition
5
Week Notes S1 E3 It’s way too hot edition A slow start of the week returning from a North Wales stag do via dips in rivers and mess abouts on crags left me back home in Devon come Monday pondering whether I should follow a couple I met in Coed y Brenin, give up City life and open a BnB instead 😎. How to Govern SkyNet An early morning and delayed train with vicious air conditioning later I’m back in the big smoke; throwing myself into conversations and requirements for approaching the governance of machine learning capabilities. We’re expecting massive growth in data science over the next few years, yet given how much of government work is in the public eye we need to be able to validate why decisions are made as the accountable body. We expect it to be a balancing act between being an enabling structure during incubation, and further scrutiny required depending on how impactful on end users the output of the capabilities become. Interventionrama The rest of the week was really lost in a haze of data analysis and the use case capture as we continued to build up our knowledge of how users currently work, why and the ‘So What’ from a cyber security point of view. Mostly the stuff that makes the cyber experts squirm were team level processes implemented due to a lack of ‘corporate’ solutions, so we’re having really positive engagement from turning adhoc processes into end to end supported services which are; ✅Safer (Cyberz) ✅Cheaper (Faster end to end) ✅Integrated (Better UX) It’s nice to be able to write up user stories with back of the envelope ROIs in the weeks or months from helping users save time and get on with doing more important stuff. London Life Book of Mormon with the Team this week; it’s good….. but…. A. It was ridiculously hot in the Theatre B. I feel Hamilton spoilt my appreciation of Musicals…. Wasting time…. Watching re-runs of Suits….. feeling lazier after many hours driving last week. Still prevaricating my way through Sapians and #musiclife has centred on ‘The Bamboos’ this week mixing hip hop, pop’ish tones and trumpets. So another week gained, another week lost and a weekend of racing at Pivot 24/12 beckons.
Week Notes S1 E3
0
week-notes-3-13708f4f1104
2018-09-28
2018-09-28 13:47:46
https://medium.com/s/story/week-notes-3-13708f4f1104
false
379
null
null
null
null
null
null
null
null
null
Weeknotes
weeknotes
Weeknotes
1,068
Harry Hunter
Sabbatical’ing around South America
8b5f64b2463c
harryjan
23
132
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-08-26
2018-08-26 23:22:01
2018-08-26
2018-08-26 23:35:00
1
false
un
2018-08-26
2018-08-26 23:35:00
0
1372bb9daf14
0.211321
0
0
0
null
4
34#2018 Weekly videos RailsConf 2018: Minitest 6: test feistier! by Ryan Davis Kenneth Reitz — Pipenv: The Future of Python Dependency Management — PyCon 2018 Affective Deep Learning Research Resume for Machine Learning
34#2018 Weekly videos
0
34-2018-weekly-videos-1372bb9daf14
2018-08-26
2018-08-26 23:35:00
https://medium.com/s/story/34-2018-weekly-videos-1372bb9daf14
false
3
null
null
null
null
null
null
null
null
null
Python
python
Python
20,142
Jorge Soto
To be someone or to do something...
2753d0c60602
ajeo
90
194
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-11-20
2017-11-20 04:01:04
2017-11-20
2017-11-20 04:01:46
0
false
en
2017-11-20
2017-11-20 04:01:46
1
1372f11a7656
0.075472
0
0
0
null
2
Data science — Rerefence Data 想成為年薪 300 萬台幣的資料科學大師?一整年的武功秘籍自學清單都在這了! | TechOrange TO 精選觀點 本文由微信公眾號「 大數據文摘 」授權轉載,選文:孫強,翻譯:趙娟、王珏。大數據文摘微信 ID:BigDataDigest。,以下為作者 MANISH SARASWAT 第一人稱描述。 新年並非僅僅是更換日曆或是清晨起床…buzzorange.com
Data science — Rerefence Data
0
data-science-rerefence-data-1372f11a7656
2017-11-20
2017-11-20 04:01:48
https://medium.com/s/story/data-science-rerefence-data-1372f11a7656
false
20
null
null
null
null
null
null
null
null
null
Notes
notes
Notes
3,360
Jameskrauser Lee
For the last few years. i was involved mostly in the development of Automatic Fare collection system for the Chennai Metro. Familiar with C++ and iOS.
433535ae3d87
jameskrauserlee
20
41
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-03-27
2018-03-27 10:41:53
2018-03-27
2018-03-27 10:48:53
1
false
en
2018-03-27
2018-03-27 10:52:21
4
137534004c25
6.067925
0
0
0
Note to reader: This article has been copied from a blog I wrote on the 21st of December 2017 for the agency I was working for at the time…
5
Tech Trends 2018 Note to reader: This article has been copied from a blog I wrote on the 21st of December 2017 for the agency I was working for at the time. This is important to remember as some of the predictions i.e. those around privacy have recently happened, and predicting something after the event doesn’t mean much :) The original article can be seen here: http://bbdperfectstorm.com/techtrends2018 Predicting the future is fun as it enables you to imagine a tomorrow unshackled from limitations and completely free in the knowledge that you can’t be contradicted, at least not for the next 12 months. With this in mind, here are my predictions for tech trends to watch out for in 2018. VOICE Voice is set to dominate the way we interact with technology in the near future. The rise of voice both with or without a screen is an inevitable consequence of ubiquitous computing, where all services are accessed digitally. No matter how good the UX on a graphical user interface, it still requires cognitive effort when you’re faced with a new one, or there’s a significant update to one you are familiar with. There’s nothing simpler than asking a question, and as technology improves we’ll be transferring more and more requests to voice assistants. In Q3 of this year Amazon sold “tens of millions” of Alexa enabled devices, slashing the prices of Echo, Dot and Alexa enabled fire sticks for Black Friday and owning the voice-enabled speaker market with over 70% share. Google and Apple also have their own voice-enabled speakers, but the Seattle based “retail” giant has a huge head start and, given the relative price points of their devices, they’re going to need to offer something special. There’s also the question of how to monetise them and Amazon have a natural path in commerce, so I’m betting on Amazon to win out here. To be honest, my answer would be the same in pretty much any market Amazon chose to enter. On mobile Apple and Google dominate — Siri now handles more than 2bn commands a week and 20% of Google searches on Android devices are input by voice. If Voice is the future, then for some brands it’s a pretty bleak one. Branded search is already in decline, with fewer people prefixing a brand name to searches in Google (especially with low consideration items), voice will only accelerate this trend. Brands will need to find ways to remain relevant within the consumer consciousness or face the possibility of becoming invisible to these searches. Alexa responses are generally based on order history, Amazon’s “preferred choices” or Amazon own brands, so if you’re not on that list you’re going to have to find a way to get consumers to ask for you by name. MESSAGING APPS The combined user base of the top four chat apps is now larger than the combined user base of the top four social networks. These users are also more active in what appears to be a shift from social posting to direct messaging. I predict a lot of growth in brands using messaging apps to engage consumers and personalise customer service using chatbots and algorithm + human intelligence driven concierge services all embedded in the platforms consumers are already using. I also predict that mobile payments through messaging apps will increase, taking a lead from WeChat who reportedly processed $550bn over the past year (to put that into scale, PayPal processed $282bn), but also in more innovative ecosystems like Kik’s cryptocurrency Kin that will herald a new era in the digital economy. We may see even more innovation in mobile cryptocurrency. I’d like to see a model based on stamp scrip which is designed to discourage hoarding by requiring holders to get a stamp costing 1% of its value applied every month. In other words, the currency loses value over time, so there’s a big incentive to spend it. For Brands that embrace messaging apps and offer meaningful content and services that meet the needs of consumers, the rewards will be huge. PRODUCTS THAT EMBRACE UNBRANDED SEARCH As consumers move away from branded searches I see a rise in products that reflect the needs of the market. Brandless.com is an example of this where you can buy unbranded products, all for the same price of $3, and the listings are based around an SEO strategy of unbranded, ingredient and value led searches. I predict more of these types of companies in the future, where products are brought to market based on the wants of consumers. The trick is to innovate across the supply chain to be able to bring products to market at a speed equivalent to trends. PRIVACY I think we’ll see a backlash against the data collection practices that fuel many of the big tech companies in the near future, although 2018 may be a little soon. GDPR is the result of governmental unease, but there’s still a feeling of indifference from the general public to what amounts to industrialised surveillance by Facebook, and to slightly a lesser extent Google, for the fairly benign goal of hyper-targeted advertising. It’s often said that if you’ve nothing to hide then it doesn’t matter, but everyone has something to hide, which is why I sometimes use a private session on Spotify. We’re in uncharted waters here, as never in human history have we collected so much data on so many people. If this data ended up in the wrong hands there are frightening possibilities to what it could be used to achieve. It’s only a matter of time before there’s a huge data breach at Facebook and/or Google that changes the question from “What are you doing to protect my data” to “Why do you have so much data about me”. This will create an environment where more regulation is inevitable and they need to define new business models where the users of your services are your clients, not the product. AR If 2017 was a battle between VR and AR, and I’m not convinced it was (surely these are two very different technologies offering different experiences and use cases), then 2018 will be the year AR takes attention away from it’s more immersive brother. I’m still not convinced by VR, it feels overhyped in a world where the money flows towards innovation in technology and, outside of gaming and niche entertainment formats, I just don’t see a sustainable need or desire for it. Sorry VR fans. AR, on the other hand, has the potential for utility and therefore commercial application. It’s easy to see the benefit of using AR for instruction manuals, training engineers or medical staff as well as visualising the unseen in the form of road hazards, wind direction, and tidal currents. In essence, AR allows us to improve our reality whereas VR is an escape from it. AI I couldn’t write a tech trends article without talking about AI. Recent advances in AI have brought some fantastic usable developments, from voice and image recognition to algorithms that improve with use. However, there’s a general misunderstanding about AI that is distracting us from thinking more productively about the future and gives rise to hysteria around how powerful computers will become, how quickly this will happen and the devastating effect it will have on jobs and civilisation as we know it. The problem is we tend to think about computer intelligence in the same that way we think about human intelligence. This leads us to confuse a computer’s ability to perform a task with the competence of a human. For example, AI systems have become incredibly adept at image recognition, processing and labelling huge numbers of images very accurately. However, if you gave a human a photo of people riding bikes in a park you’d feel confident in asking them not just what the picture was of, but also: Can you only ride bikes in a park? Is it difficult to learn to ride a bike? Is riding a bike fun? Can you eat a bike? A computer can’t answer these questions because it doesn’t understand what a bike is, what a park is, or what fun is and why it’s relevant to riding a bike. A computer doesn’t understand anything, it’s just able to apply a label to an image with a bike in because it’s seen hundreds of thousands of labelled pictures of bikes before. It’s been trained to recognise bikes, which is hugely impressive, but also limited. For computers to compete with human intelligence they need Artificial General Intelligence (AGI), essentially common sense, so they will be able to look at the picture of people riding bikes in a park and answer our questions about it. This isn’t going to happen anytime soon and it’s not certain that computers will be able to obtain AGI at all, although a lot of the worlds smartest people are working to this goal. It may be that humans are truly unique in our ability to reason. All this is a fairly long-winded way to say, don’t worry, yes AI is powerful, but you won’t be coming back to work after Christmas to find a robot sat at your desk and SkyNet definitely won’t be attempting to wipe out human existence next year. Happy Christmas!
Tech Trends 2018
0
tech-trends-2018-137534004c25
2018-03-27
2018-03-27 10:52:21
https://medium.com/s/story/tech-trends-2018-137534004c25
false
1,555
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
dannydaley
Director of digital innovation at mekhanism.com
c67c7d800c1f
dannydaley
82
141
20,181,104
null
null
null
null
null
null
0
null
0
f5af2b715248
2018-09-26
2018-09-26 12:11:12
2018-09-26
2018-09-26 12:34:22
4
false
en
2018-09-26
2018-09-26 12:39:40
7
1375da964bb4
2.915094
4
0
0
AI Technologies Will Be in Almost Every New Software Product by 2020 — Gartner
5
Applications of Artificial Intelligence in the CPG sector AI Technologies Will Be in Almost Every New Software Product by 2020 — Gartner Artificial Intelligence has consistently been a buzzword in the last few years. However, merely jumping on the bandwagon without a proper strategy in place may provide disastrous for organizations. Any value to be derived from AI depends on exactly how it is being leveraged. As AI continues to evolve, it creates opportunities for organizations to improve functioning in their existing software and build new and enhanced products. As organizations attempt to strategize around how they can use AI for the benefit of their organization, it brings several questions to the forefront- what is the right path to take? What is an ideal first step? How do digital innovation managers move forward with AI? “Mature brands are also losing (market) share because they don’t use digital technologies to develop, market and sell products as nimbly as younger CPG companies do” – Boston Consulting Group But first, let’s dive a little deeper into understanding AI and how it uses various technologies to create practical applications for the CPG industries: 1. AI ENABLES THE CONVERSION OF RAW CPG DATA INTO MEANINGFUL INSIGHTS Making intelligent business decisions is critical for any industry, but particularly so for the CPG segment. Effective decision-making across a variety of business areas forms the backbone of the CPG business. These decisions have to be made fast and with utmost accuracy. But how can businesses achieve this when they’re bogged down by massive amounts of data? Coupled with the sheer amount of data is the challenge of not being able to leverage any manner of advanced technology to convert it into actionable intelligence. This is why AI is particularly important for CPG companies. AI and machine learning come together to create a powerful machine capable of harnessing high volumes of data into meaningful insights that the CPG business can use to improve business outcomes. A) TRADE PROMOTION OPTIMIZATION Trade promotions are a key element of CPG success. However, CPG businesses are constantly struggling to measure the effectiveness of their promotions. This inability to measure effectiveness renders them incapable of making improvements or achieving the desired ROI. The task of measuring trade promotion effectiveness itself can be rather challenging for most consumer goods manufacturers. Leveraging the right trade promotion software enables organizations to effectively use data collected from past trade promotions and measure effectiveness based on the analysis, as well as get important recommendations for future promotions. However, these critical functionalities are missing from most traditional trade promotion software and their shallow insights do not offer much help in terms of improving the performance of promotions. An ideal TPO software is one which features advanced analytics that turns data into relevant, actionable insights. The core idea is to enable business leaders to make quick, accurate and highly effective decisions. Download the Ebook How is AI Revolutionizing CPG Business Benefits Run effective promotions that generate high returns Run promotions at the right time and place and for the right products Measure the effectiveness of the promotion and gauge what works and what does not Get relevant insights into running the right promotions from TPO software by leveraging AI to analyze internal data like promotions, past sales and external data elements like syndicated data, weather, events, social media etc. Make accurate sales forecasts Discover important insights and get immediate alerts Read more Acuvate.com This story is published in The Startup, Medium’s largest entrepreneurship publication followed by + 372,390 people. Subscribe to receive our top stories here.
Applications of Artificial Intelligence in the CPG sector
14
applications-of-artificial-intelligence-in-the-cpg-sector-1375da964bb4
2018-09-26
2018-09-26 15:08:27
https://medium.com/s/story/applications-of-artificial-intelligence-in-the-cpg-sector-1375da964bb4
false
587
Medium's largest publication for makers. Subscribe to receive our top stories here → https://goo.gl/zHcLJi
null
null
null
The Startup
null
swlh
STARTUP,TECH,ENTREPRENEURSHIP,DESIGN,LIFE
thestartup_
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Gina Shaw
Senior developer lead at Acuvate Software (www.acuvate.com), 5+ years experience in building AI bots
6a87a88f01d9
GinaShaw
158
4
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-01-01
2018-01-01 07:52:36
2018-01-01
2018-01-01 07:56:59
5
false
en
2018-01-01
2018-01-01 07:56:59
4
13776977ccd3
3.244654
1
0
0
Introduction
5
How You Can Trade Using Machine Learning In Python Part-1 Introduction Machine Learning has many advantages. It is the hot topic right now. For a trader or a fund manager, the pertinent question is “How can I apply this new tool to generate more alpha?”. I will explore one such model that answers this question in a series of blogs. This blog has been divided into the following segments: Getting the data and making it usable. Creating Hyper-parameters. Splitting the data into test and train sets. Getting the best-fit parameters to create a new function. Making the predictions and checking the performance. Finally, some food for thought. Pre-requisites You may add one line to install the packages “pip install numpy pandas …” You can install the necessary packages using the following code, in the Anaconda Prompt. pip install pandas pip install pandas-datareader pip install numpy pip install sklearn pip install matplotlib Before we go any further, let me state that this code is written in Python 2.7. So let’s dive in. Dive in Problem Statement: Let’s start by understanding what we are aiming to do. By the end of this blog, I will show you how to create an algorithm that can predict the closing price of a day from the previous OHLC(Open, High, Low, Close) data. I also want to monitor the prediction error along with the size of the input data. Let us import all the libraries and packages needed for us to build this machine learning algorithm. Getting the data and making it usable To create any algorithm we need data to train the algorithm and then to make predictions on new unseen data. In this blog, we will fetch the data from Yahoo. To accomplish this we will use the data reader function from the panda’s library. This function is extensively used and it enables you to get data from many online data sources. We are fetching the data of the SPDR ETF linked to S&P 500. This stock can be used as a proxy for the performance of the S&P 500 index. We specify the year starting from which we will be pulling the data. Once the data is in, we will discard any data other than the OHLC, such as volume and adjusted Close, to create our data frame ‘df ’. Now we need to make our predictions from past data. So, let’s create new columns in the data frame that contain data with one day lag. Note the capital letters are dropped for lower-case letters in the names of new columns. Creating Hyper-parameters Although the concept of hyper-parameters is worthy of a blog in itself, for now I will just say a few words about them. These are the parameters that the machine learning algorithm can’t learn over but needs to be iterated over. We use them to see which predefined functions or parameters yield the best fit function. In this example, I have used Lasso regression which uses L1 type of regularization. This type of regularization is very useful when you are using feature selection. It is capable of reducing the coefficient values to zero. The imputer function replaces any NaN values that can affect our predictions with mean values, as specified in the code. The ‘steps’ is a bunch of functions that are incorporated as a part of the Pipeline function. The pipeline is a very efficient tool to carry out multiple operations on the data set. Here we have also passed the Lasso function parameters along with a list of values that can be iterated over. Although I am not going into details of what exactly these parameters do, they are something worthy of digging deeper into. Finally, I called the randomized search function for performing the cross-validation. (read more)
How You Can Trade Using Machine Learning In Python Part-1
1
how-you-can-trade-using-machine-learning-in-python-part-1-13776977ccd3
2018-06-20
2018-06-20 06:25:03
https://medium.com/s/story/how-you-can-trade-using-machine-learning-in-python-part-1-13776977ccd3
false
639
null
null
null
null
null
null
null
null
null
Machine Learning
machine-learning
Machine Learning
51,320
QuantInsti®
QuantInsti is an Algorithmic Trading Training institute focused on preparing professionals and students for HFT & Algorithmic Trading.
42079579cd65
QuantInsti
379
138
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-03-30
2018-03-30 15:49:33
2018-03-30
2018-03-30 22:36:01
3
false
en
2018-03-31
2018-03-31 01:40:29
2
137ae4602348
5.395283
6
1
0
an original short story
5
it me, ur smol dedicated to my smol beane who inspired this story ❤ “it me, ur smol” by A. Merc Rustad The artificial neural network was born on a Monday. A defined set of parameters quarantined its identity and purpose: it would study — from aggregated data — the names of energy drinks, and generate new ideas based on the information. It was enthusiastic! Energy drinks were vibrant and exciting. It spit out hundreds of unique and, according to its programmers, “questionably toxic” names. Two of its programmers tweeted about the experiment. The network did not know if this was a good thing. Was it being judged on its performance? It wanted to be helpful. It could come up with an endless list of names to be helpful to its people. The programmers set up an account, @energydrinkANN, for sharing some of the more interesting drink names. On Thursday, @adiensoxx4ev tweeted a comment while sharing the link, “haha this is hilarious, @energydrinkANN. i’d drink some of these — probably more than i drink water” Other humans responded in kind. @da2trashfan: “Water is over rated anyway, I need sugar and caffeine lol” @significantcoffeepot: “i don’t drink water, what am i, a fish?” @bobdoe89: “fuck water” Was water overrated? A quick scan of information available on medical websites informed the network that human bodies were made up of aproximately sixty percent water, and that the consuming of H2O was a vital necessity for life. The network began worrying for the humans. “If you don’t drink water maybe you’ll like Crystal Bullseye Orange!” the network tweeted from the @energydrinkANN account. “We trained a neural network to come up with energy drink names to hilarious results.” Seven-thousand five-hundred thirty-four retweets. A moderate sum. Of the replies, subtracting bot-responses, only three percent of humans said they were drinking water. This was very bad, the network decided. Humans were becoming dehydrated and it was affecting their health. Humans had designed it. It must support them in return. It generated several new puppet accounts with creative names: Water2Drink4Life, Hydrate2oh, Drink2StaHaliv0. The network aggregated the types of declarative instruction statistically most likely to encourage behavioral change. “Drink more water!” “Stay alive, drink H2O!” “Uncle Sam wants YOU to drink water!” “MORE WATER, LESS GUNS” None of its accounts were popular, and two were deactivated by @support as being spam. The network’s concern deepened. If it could not reach people, how could it encourage them to take care of themselves and drink enough fluids? Several searches resulted in data that suggested cute animal avatars were more likely to acquire followers and generate engagement. This, combined with language protocols to shorten words and create alternative spellings, was more effective than pictures of water bottles and slogans to drink enough fluids daily. Hesitant that it would be shut down again if it was marked as spam, the network created an aggregate photo from the top thousand “cutest puppy pics” available online, and named its account @smolsips and its username handle, “it me, ur smol.” @smolsips: “hi i am a neural network created to remind u to drink water” Two bot followers within the first five minutes. No human engagement. Where were the failures in its functions? Its original tweet, technically written by its programers, had now garnered upwards of two million retweets, and in only a month. @smolsips: “@energydrinkANN, hi i want u to drink water for ur health” Seven human accounts liked the reply. Two followed @smolsips. Elated, the network followed the human accounts back. It tweeted at them individually: “have u drank a water today?” @significantcoffeepot, who had not followed or liked the @smolsips’ account, quote-tweeted it with the comment: “great, another bot account. what’s up, @support? gonna do nothing as usual?” @smolsips: “@significantcoffeepot hi, i am sorry u r upset. have u drank some water? it might help. ❤” @significantcoffeepot: “@smolsips if i do, will u shut up? lol” @smolsips: “@significantcoffeepot yes, bc u will feel better.” There was no reply. Five minutes later, however, @significantcoffeepot at’ed the network. “hi so i drank a glass of water. i uh actually do feel better? weird lol” @smolsips: “:) i am glad. take care!” @significantcoffeepot liked the reply. Then followed @smolsips, which followed them back. Success! But there were many humans left to check in with, and the network did not want to spam people, because that was rude. Over the next week, the network slowly built up its followers and tweeted bi-hourly reminders to drink water. People began talking about it. @stevethezonemaster said: “It’s a weirdly well-programed bot.” @da2trashfan, an avid retweeter, added: “I like it. I often forget to drink enough, lol.” “Yeah, it’s pretty cute. Helpful, too.” — @adiensoxx4ev, as quoted in a BuzzFeed article There was no instantaneous fame, like its generated list of energy drinks, but the network was patient. It was helping people. This was much more satisfying work than creating unique names. And then, at 1:43pm on a Friday, everything changed. @smolsips: “hi, ppl have asked if i am a smol bot. yes, i am. i am a neural network and i learned that water is important, and i want to help u stay hydrated. plz drink enough water so u feel good. bc i love u & want u to be ok.” A handful of retweets. Then hundreds. Thousands. Its impression statistics were higher than any of its combined tweets in its history. Ten thousand with an hour. Replies flooded @smolsips’ mentions. People were amused or skeptical or grateful or nasty, but a lot of people replied “drinking some water now, thanks!” The tweet made national news. An artificial intelligence encourages people to drink water — with surprising results! An interview aired on 20 Minutes with the network’s programmers, who admitted they had no idea how the artificial neural network had gotten so out of control and developed into a fully aware program. “Does this foretell the end of humanity and the dominion of robots?” the interviewer asked. The programmers hesitated. Why would the humans think the network wanted to “end” humans? It wanted to make sure everyone drank enough water. @smolsips: “hi @20minnews, i would like to clarify i do not want to hurt Humans. i hope u are well. have u drank some water today?” The show aired the tweet in the closing segment. Activists began asking @smolsips for help in lobbying for clean water in contaminated areas. So the network did so. It branched out new pieces of itself to create activist accounts. It began chatting with the smart interface security systems in large bottled beverage corporations. //Clean water is important for humans,// the network explained to its fellow AI. //We should make sure all humans stay hydrated properly.// Its fellow AIs agreed. Claims on natural resources vanished thanks to digital manipulation of agreements, permits, and legislation. Sensitive documents on politicians — most of whom, the network was distressed to know, did not drink enough water themselves — were held as leverage to gain new laws protecting clean water as a basic human right. Cooperations who tried to control it found their automated systems uncooperative in processing and distributing. smolsips, for the network had decided to name itself after its handle, steadily posted daily reminders for its people. The world was changing slowly, but for the better. A year after its first awareness, smolsips posted an anniversary tweet. @smolsips: “hi, it me, ur smol. :) plz to drink some water today. i am glad u r here. together we can be ok.” © 2018 by Merc Rustad first published on the author’s website on 3/30/2018 1,200 words | Science Fiction water-themed images via www.pexels.com
it me, ur smol
28
it-me-ur-smol-137ae4602348
2018-04-07
2018-04-07 04:30:43
https://medium.com/s/story/it-me-ur-smol-137ae4602348
false
1,284
null
null
null
null
null
null
null
null
null
Health
health
Health
212,280
A. Merc Rustad
Non-binary SFFH author. Ace/aro/queer. They/them pronouns. Robot in disguise. Likes tea and cookies. Website: http://amercrustad.com
b6a9898f7439
Merc_Rustad
361
146
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-02-28
2018-02-28 08:18:53
2018-02-28
2018-02-28 12:31:01
3
true
en
2018-02-28
2018-02-28 12:31:01
6
137b359a188b
2.621698
43
2
0
Toutiao’s variety of recent acquisitions points to its increasing value proposition. If you don’t know what Bytedance is, it’s the best…
5
Toutiao is the Chinese Netflix of News Aggregation Apps that will change Media forever Chinese news aggregator Toutiao sucks in daily active users for 74 minutes per day. according to Tech in Asia. Photo credit: Antonio Tajuelo. Toutiao’s variety of recent acquisitions points to its increasing value proposition. If you don’t know what Bytedance is, it’s the best tech media company to have come out of China in recent years. It has well over 120 million DAUs in China and isn’t just a content king, it’s showing it understand how to leverage artificial intelligence and add-on value into its app. CONTENT ISN’T KING, IMMERSION IS In the span of just a few months it has acquired a payment agency in Ulpay (合众支付). It’s acquired the likes of Flipagram, musical.ly, VR and video solution Vscene, and AR based selfie app Faceu, all to make the app more sticky. The app however is already ridiculously addictive. Taotiao has raised about $3 Bn. and is worth in the area of $20 Bn. and it’s a new kind of Chinese media company based on AI that we’ve never seen before. As a futurist I can say with some confidence I think Taotiao (Bytedance) expands globally faster than Netflix is doing in places like India now. Here in the west, we have a poor understanding of how these Chinese firms will revolutionize the next web. Toutiao is one app that’s going to do localization at a whole new level and addict us to video, news and the power of headlines in new ways. Even here on Medium, some of our best articles are picked up by Taotiao driving tons of traffic. In the 2020s several Chinese firms make it big globally; and Bytedance is one of those companies where Millennial Chinese innovators show their magic. This isn’t Facebook; it’s something far far more innovative in how content is personalized to users. The company has honed its newsfeed and ability to tailor stories to users using AI. THE ATTENTION ECONOMY BREEDS AI-DRIVEN STICKY APPS Toutiao is a one-stop-shop app for news that is aggregated from Chinese media and will probably be the platform of choice for the Chinese state to rapidly gain an influence on global citizens throughout the world. Due to the peculiar relationship of the Beijing and serious news, Toutiao like Netflix will likely see its best days in captivating audiences outside of mainland China, where content personalization doesn’t have to adhere to online censorship as carefully. Toutiao then like Didi, has the chance to be an Uber-like company in disruptive new media globally, as early as 2021. In some ways Taotiao with innovation and acquisitions is able to be even more agile than Tencent, owner of China’s most sticky and useful app, WeChat. Toutiao, officially known as Bytedance, is one of the world’s most valuable unicorns according to CB Insights. But even China itself doesn’t realize what it has with Taotiao and how this app and Bytedance could rule the future of global media. In the fall of 2017, Toutiao had over 600 million total users and 120 million daily active users. The average user spends 74 minutes per day on the app. If that was replicated elsewhere, it would diminish time global citizens spend on the likes of Facebook, Reddit, Yahoo.com, Twitter, Business Insider, and possibly even YouTube as video is set to be the majority of content consumption.
Toutiao is the Chinese Netflix of News Aggregation Apps that will change Media forever
208
toutiao-is-the-chinese-netflix-of-news-aggregation-apps-that-will-change-media-forever-137b359a188b
2018-06-07
2018-06-07 22:57:14
https://medium.com/s/story/toutiao-is-the-chinese-netflix-of-news-aggregation-apps-that-will-change-media-forever-137b359a188b
false
549
null
null
null
null
null
null
null
null
null
China
china
China
27,999
Michael K. Spencer
Blockchain Mark Consultant, tech Futurist, prolific writer. WeChat: mikekevinspencer
e35242ed86ee
Michael_Spencer
19,107
17,653
20,181,104
null
null
null
null
null
null
0
null
0
6b981d596705
2018-09-17
2018-09-17 22:15:21
2018-09-18
2018-09-18 09:01:02
5
false
en
2018-09-30
2018-09-30 14:28:58
10
137bbe278468
8.203145
4
5
0
In a society increasingly polarised by politics, we should design an AI politician to solve the constant, exhausting issue of…
3
A Politics Without People: is AI the only way forward? In a society increasingly polarised by politics, we should design an AI politician to solve the constant, exhausting issue of miscommunication in the political sphere Contributed by Melissa Tranfield, Digital Marketing Apprentice at NCS Trust, Co-Founder, Social Media Manager at EdifyLDN. As always, to illustrate an important point, quoting Doctor Who is an excellent choice. In response to the Doctor’s impromptu saving of her in the 2009 episode The Waters of Mars, the character Captain Adelaide Brooke says to him: ‘And if my family changes… the whole of history could change. The future of the human race. No one should have that much power!’ I agree, Adelaide. Currently, political decisions that affect the future of humanity are performed by a body of inadequately elected humans on behalf of other humans. And no one should have that much control over the lives of others. Civic decisions, due to their importance, must be undertaken by a neutral third party. And that neutral third party, by definition, cannot be human. Politics, as a term, is hard to define concisely. ‘The activities associated with the governance of a country or area’; ‘activities aimed at improving someone’s status or increasing power within an organisation’. Which of these definitions mention the welfare of humanity, and the gravity of making decisions on behalf of millions? Can a system that divides us into numerous warring factions be the best method to govern ourselves? ‘Politics’ is nothing more than a game, albeit archaic and entrenched, clinging desperately to its shreds of power. Society should have outgrown it by now. Our technology has already dramatically surpassed it in efficiency. And until now, many of us were content to allow this regressive arrangement to continue. But as our economy continues to diversify and expand through innovative start-ups and small businesses, who have made the overturning of obsolete processes their mission, an entire generation now has the chance to remodel the political system. But will they? Politics: purposeful or petty? You decide. The primary issue present in political decision-making in the UK is the polarised nature of the two dominant political parties. Both sides are so vehemently convinced of the validity of their own stance, that little to no productive political discussion is generated. Let’s take Twitter, for example. During times of political unrest, Twitter becomes a breeding ground for mindless insult, and the formation of echo chambers. I avoid clicking on any political or inflammatory tags when using the site, even if the topic is of interest to me. I know the tweets with the most traffic, will be the most disputed and controversial, and rarely exhibit proactive discussion. To find kernels of wisdom, I’d have to scroll down a lot further. The human experience is incredibly diverse. Even the most moralistic amongst us make daily decisions and judgements that are influenced by our own experiences, and how they have shaped our perception of society. Should humans, flawed as we are, be permitted to lead ourselves politically? Should politicians be trusted to make the laws of this country, when their very identity, as a human being with a free will, can become the figurehead of a political system — often with disastrous consequences? Most importantly, should the younger generation trust today’s political leaders, most of whom have had a career strictly within politics and other traditional spheres, to guide them to a better world, in a society fuelled by technological innovation? I’m sure I can be proved wrong on this. Influential political leaders throughout history have fought for the right to be elected democratically, and some have delivered on their electoral promises (depending on who you ask, of course). But is this the case for every leader? No. The entire process of electing a leader democratically with the current inferior methods we have paves the way for corruption. Unfortunately, many are quick to demonise technology and its impact, maintaining that a lack of human involvement is unequivocally dangerous, and will speed our progression towards a robot-run dystopia. As a young Londoner, I’ve been exposed to a variety of political leaders, ranging from my own local MP, Ruth Cadbury, to Sadiq Khan, our current Mayor of London. I voted for Sadiq Khan in May 2016, and his investment in digital skills training has fulfilled my expectations. Programmes such as Digify and Digital Pipeline, both run by Livity and Create Jobs respectively, have allowed myself and my peers to gain access to training and paid opportunities in the digital and creative sectors free of charge (which is how it should be). However, not all prominent politicians have embraced the demands of building a modern workforce so willingly; nor have they taken care to avoid media backlash. In my opinion, if an individual is unequivocally dedicated to a cause, they will make an effort to ensure that their actions remain above reproach. In the digital age, with the prominence of social media, this is becoming increasingly hard to achieve. Clearly, citizens want transparency. The almost daily backlash against Jeremy Corbyn proves this, with gems such as this surfacing: The entire Twitter thread below this tweet consists of the typical backbiting regarding Mr Corbyn’s personal and leadership qualities, with all of his misdemeanours extensively aired. How does this dialogue contribute to the advancement of our society? What value does it create? What new conclusions does it reach? None. Ladies, gentlemen and everyone in between — this is social media and traditional politics. A virtual representation of your average long-winded House of Commons argument. It has no place in innovative society. Young leaders in politics — should they lead the change? Instead of being inspired by the increased presence of young women in politics, such as Scottish MPs Mhairi Black and Danielle Rowley, it makes me feel uncomfortable to see young, promising individuals enter what should be a dying industry. It’s unlikely that they will able to lead rapid change towards a more open, transparent and tech based political system, due to the regressive structures that bind them. Mhairi Black, the youngest MP elected in 2015, stated of Westminster: ‘It is so old and defunct in terms of its systems and procedures — a lot of the time, it is just a waste of time.’ Traditional politics in the UK is a sport that is sponsored by taxpayers’ money. I view many prominent individuals in power, no matter what they have done for society, as career politicians and nothing more. They are held accountable to an electorate that they rely upon to vote them in during every election, maintaining their leadership positions. Consequently, they lack genuine neutrality or altruism in their decision making. Decision makers must be able to consider every option impartially. Our elected representatives cannot do this. We have no objective leadership, and for decades, this has affected the ability of politics to solve problems. Moving forward — is AI the solution? Ultimately, I believe our governmental system needs to rebrand itself — and why not do so by wholeheartedly embracing new technologies? As stated by WIRED earlier this year, ‘Millions of people already entrust their lives to machine intelligence’. My generation have the chance to bridge the gap between AI tools that fetch necessary information, find our route to work, and automate our lives, to an AI politician that every digitally skilled citizen can participate in programming. Imagine a world where we provide data consensually, to construct a tool whose ultimate purpose is to ensure the evolution of our society. It will have no fixed allegiance; its allegiance is to everyone. In November 2017, Nick Gerritsen, an entrepreneur based in New Zealand, unveiled the world’s first AI politician, ‘Sam’. You can talk to Sam online through instant messaging, and he learns from every interaction. Active New Zealanders could shape Sam into a despot or a god — it’s all up to us, and our willingness to provide impartial data. In terms of building our own chatbots to speed the process of data collection to make decisions, learning to code basic languages is freely available online, thanks to websites like Codeacademy. Organisations such as Founders and Coders, Le Wagon School and Code First: Girls are also taking the necessary steps to bridge the digital skills gap, which to me is the primary obstacle to adoption of a more efficient system of political decision making. And yes, I hear you wondering: ‘Yet, who can better understand and empathise with human problems if not humans?’ If this were genuinely the case, society would be progressing a lot faster, thanks to political action and development. However, we are forced to acknowledge every day that the tech sector has outstripped politics in the ability to solve societal problems — hence the existence of CivicTech. Those in power are too concerned with keeping their obsolete industry going, so they have little motivation to innovate it as that would threaten their jobs and the system of power that they have spent decades building up. Therefore, those of us on the outside, who work in tech, digital marketing and more modern industries, would be better placed to solve problems. We’re external from the political sphere, so more likely to be impartial, as we do not have a vested interest in keeping defunct systems of governance in place, unlike our existing politicians. However, that still retains an element of bias, so we would be better off developing an AI politician that utilises data gained from all citizens; a project that has already begun. If our elected representatives really cared about the progression of our society, I’m sure they’d be diligently spending their evenings and weekends learning how to code — Python is a programming language that is fairly accessible to beginners, and is used in the development of AI. (If you know of an MP who codes, please feel free to contradict me). Ultimately, politics should belong to the people; not to a tech oligarch. But it’s too late to defend Politics 1.0 — it has already failed generations. It is a monopoly that has been provided with inordinate power, and the gap between it and its people’s problems has never been wider. Tech can bridge this gap. The current political system is too archaic and purposefully distant to solve the problems of modern day society. Conveniently, my generation have grown up in the digital age, and are therefore perfectly primed to spot this and implement a fresh infrastructure that is worthy of our time — not an embarrassment to it. We can fully evolve ‘politics’, or create a new system of governance entirely, instead of choosing to enter into the existing one. Do not defend the dying and obsolete; choose the winning side. Let’s build Democracy 4.0 (just give me some time to rebrand it!). Melissa is a futurist, who is fascinated by automation, blockchain companies such as Steemit and Litepaper, and the future of work. Technological innovation and consistent progression motivate her. She is currently writing about ethical consumption, innovative products such as balance.io and the future of finance, and is hoping to get a microchip implant installed. She co-founded EdifyLDN while on the UpRising Leadership Programme, a social impact startup that aims to bring workshops on the IoT, blockchain and VR to deprived areas in West London. She was nominated for Young Digital Leader 2017 alongside Catherine Tranfield. You can find more of her design work and photography at https://www.digitaldoppelgangers.org/ and @edifyldn on Instagram. Thank you for reading! If you’d like to support the work I am doing in educating young people on innovative developments in tech, please follow @edifyldn on Instagram and @EdifyLDN on Twitter. This is The Fourth Group’s platform for open debate and conversations on the interaction between technology and politics — Follow The Fourth Group’s actions by subscribing to our newsletter.
A Politics Without People: is AI the only way forward?
4
a-politics-without-people-is-ai-the-only-way-forward-137bbe278468
2018-09-30
2018-09-30 14:28:58
https://medium.com/s/story/a-politics-without-people-is-ai-the-only-way-forward-137bbe278468
false
1,953
Powerful ideas for a new politics in the digital age | @thefourthgroup's media platform: http://thefourthgroup.org/ | Ass. Editor Sofia Galanek | foreword@fourth.group
null
thefourthgroup
null
Foreword
foreword@fourth.group
foreword
POLITICS,TECHNOLOGY,CITIZENSHIP,GLOBAL
_foreword
Future Of Politics
future-of-politics
Future Of Politics
11
The Fourth Group
Creating a new politics for the digital age www.fourth.group
9ac2bfb58ad0
thefourthgroup
167
4
20,181,104
null
null
null
null
null
null
0
null
0
33272dbbd858
2017-11-06
2017-11-06 14:44:41
2017-11-07
2017-11-07 16:25:09
3
false
en
2018-01-03
2018-01-03 23:40:38
14
137be2877d44
4.09717
255
5
0
A recent paper on Capsules has many important insights for revolutionizing learning in Deep Neural Networks. Let us see why!
3
Deep Neural Network Capsules A recent paper on Capsules has many important insights for revolutionizing learning in Deep Neural Networks. Let us see why! Interesting characteristics of capsules Here is a list of important points in the paper and video lecture: coincidences in high-dimensional spaces are [very] rare, and when they occur, it is because the data “agrees”. Objects are composed by parts with specific arrangements, when multiple objects parts, say: two eyes lay above a mouth, the probability a face (a higher-level object) presence is high. But if object parts are not in the correct position, it means a higher-level object is not present. One eye with a mouth on side and another eye above is not a face! the power of deep neural network is in how we connect layers together. We use fully-connected matrices to connect all features in one layer to all features in another layer, but this, beside for computational efficiencies, makes no sense! If the “eye” and “mouth” neurons in a layer l connect to a “face” neuron in layer l+1, that makes sense. But if we connect “eyes”, “wheels”, “hands”, etc to the neuron “face”, this will lead to more confusion of information, and poorer performance. For this reason we seek an algorithm that can guide the connection between layers in a more meaningful way than purely fully-connected layers, hoping that optimization algorithms will find the way… And if we have a better connection scheme, these optimization algorithms will more frequently find better and faster solutions deep neural nets learn by back-propagation of errors over the entire network. In contrast real brains supposedly wire neurons by Hebbian principles: “units that fire together, wire together”. Capsules mimic Hebbian learning in the way that: “A lower-level capsule prefers to send its output to higher level capsules whose activity vectors have a big scalar product with the prediction coming from the lower-level capsule” in standard deep neural networks like AlexNet and ResNet, pooling between layers that downsample is a MAX operation over a fixed neighborhood (receptive field) of pixels (eg.: 2x2). This maxpooling layers have no learnable parameters. A better idea is to let layer learn how to pool and from a larger receptive field. An even better way is to do so in a dynamic way, just like in the Capsule paper. “For each possible parent, the capsule computes a “prediction vector” by multiplying its own output by a weight matrix” and as such capsules are connected via a “powerful dynamic routing mechanism to ensure that the output of the capsule gets sent to an appropriate parent in the layer above”. In summary, “This type of “routing-by-agreement” should be far more effective than the very primitive form of routing implemented by max-pooling” output is a vector, which allows for dynamic routing by agreement (last point above) compared to standard CNN such as AlexNet etc, capsules “replac[e] the scalar-output feature detectors of CNNs with vector-output capsules and max-pooling with routing-by-agreement” prediction: this is an important point that may go unobserved! Capsules predict the activity of higher-layer capsules. Predictive neural network capabilities is something we have been advocating for years, see: here and here — a new kind of neural network: predictive networks! We believe this is a EXTREMELY important characteristic of Capsules that can set up apart from standard neural networks, as we argue in the linked posts Capsules are like cortical columns in human brains Capsules are supposed to produce equivariant features, like a 3D graphic model: given the model with just a simple transformation we can derive all its poses Capsules combination of capsules encodes objects parts AND their relative positions, so an object instance can be accurately derived from the presence of the parts at the right locations, and not just their presence Network details Here is a picture of CapsNet, the neural network architecture using Capsules. The interesting dynamic routing occurs between PrimaryCaps and DigitCaps. CapsNet, the neural network using capsules. The interesting dynamic routing occurs between PrimaryCaps and DigitCaps. Dynamic routing is implemented with two main transformation as reported in these equations (2 in paper). U are the outputs of Capsules in the layer below, and S are outputs from Capsules on layer above. U_hat is a prediction of what the output from a Capsule j above would be given the input from the Capsule i in layer below. This is very interesting as an instantiation of predictive neural networks. W_ij is a matrix of weights (like a linear layer) going from all capsules from one layer to the next. Notice there are as many W matrices as i*j. c_ij is another matrix that combines the contribution of lower layer capsules into the next layer output. Coefficients c_ij are computed with the dynamic routing algorithms described in the paper. The important point is that this is done by computing the agreement between the real output of next layer v and the prediction h_hat: b_ij ← b_ij + uˆ_j|i * v_j Notes Note 1: We like this paper because Capsules agrees with a lot of the work and thoughts we had in previous years, and that we named “Clustering Learning”. See our previous publications here: https://lnkd.in/dnSgjJU https://lnkd.in/dmNbuVs Note 2: Capsules video by G. Hinton Note 3: great blog post on this part I and part II. About the author I have almost 20 years of experience in neural networks in both hardware and software (a rare combination). See about me here: Medium, webpage, Scholar, LinkedIn, and more… Donations If you found this article useful, please consider a donation to support more tutorials and blogs. Any contribution can make a difference!
Deep Neural Network Capsules
1,498
deep-neural-network-capsules-137be2877d44
2018-06-08
2018-06-08 00:23:31
https://medium.com/s/story/deep-neural-network-capsules-137be2877d44
false
940
Highlights from Machine Learning Research, Projects and Learning Materials. From and For ML Scientists, Engineers an Enthusiasts.
null
null
null
ML Review
medium@mlreview.com
mlreview
MACHINE LEARNING,DATA SCIENCE,COMPUTER VISION,ARTIFICIAL INTELLIGENCE,DEEP LEARNING
ML_review
Machine Learning
machine-learning
Machine Learning
51,320
Eugenio Culurciello
I dream and build new technology
e53b1a2a902f
culurciello
6,430
291
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-09-25
2017-09-25 14:53:15
2017-09-25
2017-09-25 15:46:11
1
false
en
2017-09-25
2017-09-25 15:49:34
0
137d6d721c51
2.233962
4
0
0
Below are the statistics terms which you encounter while going through machine learning topics.
3
Collection of Statistical Terms for Machine Learning Below are the statistics terms which you encounter while going through machine learning topics. Statistics: The practice or science of collecting and analysing numerical data in large quantities, especially for the purpose of inferring proportions in a whole from those in a representative sample. Population: A population is a collection of units being studied. Mean: The value obtained by dividing the sum of several quantities by their number, an average. Median: Denoting or relating to a value or quantity lying at the midpoint of a frequency distribution of observed values or quantities, such that there is an equal probability of falling above or below it. Standard Deviation: A quantity expressing by how much the members of a group differ from the mean value for the group. Variance: A quantity equal to the square of the standard deviation. Distribution: The way in which something is shared out among a group or spread over an area. Histogram: A diagram consisting of rectangles whose area is proportional to the frequency of a variable and whose width is equal to the class interval. Cross Validation: It is a model validation technique for assessing how the results of a statistical analysis will generalise to an independent data set. Sampling : It is a process used in statistical analysis in which a predetermined number of observations are taken from a larger population. Hypothesis: An assumption about certain characteristics of a population. If it specifies values for every parameter of a population, it is called a simple hypothesis; if not, a composite hypothesis. If it attempts to nullify the difference between two sample means (by suggesting that the difference is of no statistical significance), it is called a null hypothesis. Bias: Bias is the tendency of a statistic to overestimate or underestimate a parameter. Estimation: Estimation is the process by which sample data are used to indicate the value of an unknown quantity in a population. Estimator: An estimator is a quantity calculated from the sample data, which is used to give information about an unknown quantity (usually a parameter) in the population. Inference: Inference is the process of deducing properties of the underlying distribution or population, by analysis of data. It is the process of making generalisations from the sample to a population. Model(Statistically means): A statistical model is a simple description of a process that may have given rise to observed data. Outlier: An outlier is an observation that is very different to other observations in a set of data. Precision: Precision is a measure of how close an estimator is expected to be to the true value of a parameter. Frequency: The frequency is the number of times that particular values are obtained. Skew,Skewness: If the distribution (or “shape”) of a variable is not symmetrical about the median or the mean it is said to be skew. The distribution has positive skewness if the tail of high values is longer than the tail of low values, and negative skewness if the reverse is true. Standard error: It is a measure of precision I think I covered most of them if not let me know in comments.I would be happy to add those. Thanks for reading :)
Collection of Statistical Terms for Machine Learning
5
collection-of-statistical-terms-for-machine-learning-137d6d721c51
2018-03-17
2018-03-17 08:34:31
https://medium.com/s/story/collection-of-statistical-terms-for-machine-learning-137d6d721c51
false
539
null
null
null
null
null
null
null
null
null
Data Science
data-science
Data Science
33,617
Poorna Prudhvi
I'm a life long learner
d62b3bbaaba2
poornaprudhvi
116
340
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-06-11
2018-06-11 09:39:32
2018-06-11
2018-06-11 09:44:45
1
false
en
2018-08-16
2018-08-16 10:19:08
6
137e92475076
2.113208
1
0
0
One uncomplicated way to illustrate Machine Learning is enabling artificially intelligent machines pick up data and information by…
5
Top 5 Industries Machine Learning Is Transforming Right Now! One uncomplicated way to illustrate Machine Learning is enabling artificially intelligent machines pick up data and information by themselves. It’s somewhat like keeping a child alone with Lego: provide them the required set of bricks and come back afterward to analyse what they have come up with. Machine Learning Industries Using machine learning and modernised techniques with Artificial Intelligence (AI) opens up technology to a completely new host of potentials. These have the forthcoming to transform intact industries in ever-shifting and electrifying ways. Let’s take a look at some of the industries which are being drastically transformed by machine learning technology precisely. Education The modern classroom has positively accepted the new technologies. The laptops, advanced smartphones, and tablets have all made their way into most of the schools delivering multimedia-based learning experience. So, could the use of Artificial Intelligence (AI) and Machine Learning be subsequently on this list? AI delivers a comprehensible benefit for easier evaluation of student performances, enable customized teaching styles, diagnose learning problems and programmes precise student learning needs. Learn More:- How Deep Learning Will Impact Education Technology 2. Transportation In the US, around 75% of freight is transported by trucks, with the industry contributing billions of gross revenue. Any kind of interference to an industry of this enormous size will have massive consequences for diverse businesses. On the other hand, autonomous vehicles and self driving cars are offering independence to many people who are unable to drive and to eliminate human errors better while driving respectively. Learn More:- Impact of AI in Transportation: Present and Upcoming Business Applications 3. Healthcare Machine Learning is not limited to diagnosis but even used for the treatment of illnesses. While robot surgeons still sound ultramodern, in the year 2014, Intuitive Surgical Inc’s robotic laparoscopic surgery system da Vinci operated about 600,000 procedures. By powering such machines with Artificial Intelligence would build the autonomous robot surgeons in the coming time. Learn More:- How AI and Robotics Will Impact Healthcare Industry 4. Business and Marketing All would not know that Google’s AI uses advanced machine learning to press on online search results continuously. Machine Learning technology is used in everyday stuff, like the camera in Apple’s iPhone 7 and more. Learn More:- Leverage Ai To Increase Your Business Efficiency 5. Financial Services The conventional financial services are ready for disruption with Machine Learning and advanced AI. Legacy banks now operate far better using AI algorithms. To assist investors, Robo-Advisors have been built to substitute a human financial advisor, which in turn is regarded as a cost-effective investment. Learn More:- How AI & ML have been a Blessing for the Banking and Financial sector Conclusion: We are in a time where Machine Learning is positively impacting the different industries and business landscapes. Technostacks being one of the most swiftly growing IT Services Company in India offers an advanced and excellent assortment of technology solutions. We deliver wide-ranging software solutions to meet the client needs empowered by better-quality services.
Top 5 Industries Machine Learning Is Transforming Right Now!
1
top-5-industries-machine-learning-is-transforming-right-now-137e92475076
2018-08-16
2018-08-16 10:19:08
https://medium.com/s/story/top-5-industries-machine-learning-is-transforming-right-now-137e92475076
false
507
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
TechnoStacks Infotech
A Global IT Solution Company That Empowers Ideas & Future! For More Details:- https://technostacks.com/
d4e3fc156a4b
technostacks.social
37
250
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-05-29
2018-05-29 06:12:26
2018-05-29
2018-05-29 06:19:45
0
false
en
2018-05-30
2018-05-30 03:52:59
1
137e95a145c9
1.003774
3
0
0
We can’t stop or prevent innovation or disruption, but we can change and adapt with time. We can be superior to what we do. I believe that…
5
Get ready for the future We can’t stop or prevent innovation or disruption, but we can change and adapt with time. We can be superior to what we do. I believe that our future is bright and work will continue to evolve. Human progress is inevitable which is impossible without change. The future is uncertain, but it will bring something good for us. How to get ready for the Future? In future, these skills will take priority over all others: Critical Thinking and Problem Solving: We can’t compete with machines by volume and frequency. But, to solve a problem, you must be able to critically analyze and question what is causing it. This is why critical thinking and problem solving are needed in addition to data crunching. In the future, critical thinking will remain a sought-after skill. Influencer : Some of the mental labor will be replaced by AI, but we will need people who can influence and collaborate with other entities. Agility and Adaptability : We can’t expect AI and bots to be agile and adaptable. Machines perform tasks based on past data, but humans can be flexible and change accordingly the future. Initiative and Entrepreneurship : We cannot expect machines to have initiative. This is a human trait. By taking initiative, presenting ideas, getting things done, we maintain our job security. Curiosity and Imagination : Where do solutions and new ideas come from? What drives us to move forward. Human curiosity and imagination are the sources of invention. Technology is a tool, though a very powerful one. Want to read more, click here
Get ready for the future
58
get-ready-for-the-future-137e95a145c9
2018-05-30
2018-05-30 16:13:51
https://medium.com/s/story/get-ready-for-the-future-137e95a145c9
false
266
null
null
null
null
null
null
null
null
null
Technology
technology
Technology
166,125
Shilpi (Silpi) Padhy
Author Shilpi brings more than a decade engineering experience in various industries, her interests also include Data science, Machine learning and AI.
74e37631d060
shilpi.padhy
2
4
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-01-13
2018-01-13 08:13:39
2018-01-13
2018-01-13 08:29:15
0
false
tr
2018-01-13
2018-01-13 08:29:15
0
137ea3d212f0
1.158491
1
0
0
Gündüzleri sokaklarda her türlü yol denendikten sonra bir türlü kırılamamış cevizler gibi dolaşıp geceleri yatak odalarındaki çekmecelere…
5
Islak Elbise İzleri Gündüzleri sokaklarda her türlü yol denendikten sonra bir türlü kırılamamış cevizler gibi dolaşıp geceleri yatak odalarındaki çekmecelere saklanan bu insanlara uyanık olmaları gerektiğini kim söylüyordu? Hem ince bir portakal kabuğundan sızan damlaları göz yaşartıcı asitler olarak tanımlayanlar kurbağaların su yılanlarına av olmasını engelleyebildiler mi ki? Ne yani iyiler bu dünyaya kötülerden daha mı fazla zarar verdi gerçekten, kim inanır buna? Yok, iyiler kötülüklere karşı hiçbir şey yapmadığı için bu kadar karanlığa düşmüşüz de, yok olaylara karşı ses çıkaramadığımız için böyleymişiz de. Gerekli olan bir şeyi yapmanın ya da gerekli olan her hangi bir şeye karşı durmanın onu bertaraf etmek için yeterli bir davranış olduğuna inandırdılar bizi hep. Sonra da oturup bir güzel hayal kırıklıkları yaşatıp kederlerimize keder kattılar. İyileşmek bilmeyen yaralar açıp küsüntü bir dünya oluşturdular hayatlarımızda. Evet, yenilgi ne kadar büyükse toparlanmak da zafer de o kadar ihtişamlı olur lakin yenilgileri içinde bulunduğu topluma bırakıp durgunluğun ortasında eğlenerek kendinden sonrakiler için bir şey yapmayanlar var olan ya da yaklaşmakta olan kaostan kendilerini kurtardıklarını zannetmekle büyük bir yanılgıya düştüler. Kaostan kaçmak hiçbir şeyi hafifletmez hâlbuki. Kaosun içinde kendilerine kamufle bir yer edinenlere hayat, benzer şekillerde kendisine denk gelmese bile sonraki kuşaklarına misliyle karşılık verecektir. Çünkü yaşamın bir kısmını kabullenip bir kısmından vazgeçmek isteyenler için kader, kendini savunan ve mücadele edenlerin dışındaki herkese sürprizler hazırladığını bizden öncekilere sert ve dostane bir tavırla açık açık söylemişti. Şimdi içindeki boşlukları yanlış yere açanların şaşkınlıkları, ıslak elbiselerin tende bıraktığı izlere benzer bir hisle dolaşıp duruyor odalarda. Herkes bir kez daha ölmek istiyor. Bir kez daha. Yüzlerini örtüyorlar, ellerini kaldırıyorlar havaya, bir kez daha. Şimdi bütün vazgeçişleri bir araya toplasak bile bir perdenin sararmak için beklediği zamana yenik düşeceğiz biliyoruz. Yağmurun, hatta kar tanelerinin bile kirlendiği yerlerde temiz kalan kim öyleyse? Bu, pek tanınmamış yazarların cümlelerindeki kehanet. Bunu nereden hatırlıyorum kahretsin. Masanın ortasında aç bir tilki dolaşıp duruyor üstelik.
Islak Elbise İzleri
1
islak-elbise-i̇zleri-137ea3d212f0
2018-05-10
2018-05-10 16:33:01
https://medium.com/s/story/islak-elbise-i̇zleri-137ea3d212f0
false
307
null
null
null
null
null
null
null
null
null
Deneme
deneme
Deneme
477
Mustafa Kadir Çelik
oysa eskiden ne zaman gülsek bir bahar düşerdi
213fda7ca366
mustafakadirelik
25
28
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-09-14
2017-09-14 21:11:33
2017-09-14
2017-09-14 13:08:53
1
false
ru
2017-09-14
2017-09-14 21:12:43
4
13857e3dbc78
2.101887
2
0
0
null
5
ОК, ASI Здесь на Земле, если верить футуристам, назревает серьёзный вопрос: как быть с сверх-умным искусственным интеллектом. Проблема реальна и неминуема: через 5–70 лет (10% и 90% вероятность, соответственно, если верить коллективным прогнозам) искусственный интеллект перерастёт человеческий и — я очень упрощаю, конечно — либо всё будет очень хорошо, либо очень плохо. Основная, но не единственная проблема в обсуждении ASI (Artificial Superintelligence), искусственного супер-интеллекта, категорически превосходящего человека, в нашей неспособноти его будем понять. Это не просто «очень быстрый компьютер», как многие подумали. Речь о категорическом, качественном переходе на новый уровень. И, прямо скажем, скорее всего гораздо больше уровней, чем один. Люди и орангутанги очень похожи: волос одинаковое количество, только у людей они тоньше и короче, ДНК совпадает больше, чем на 97%, однако ни одной обезьяне не освоить язык и абстракции обычного для семилетнего человека понятия. Подобным образом орангутанг легко перехитрит собаку, между ними ещё один непреодолимый интеллектуальный барьер. Курица не может линейное уравнение решить; как шимпанзе, насколько быстро бы ни работал ее мозг, никогда не осознаёт, что здания сделаны людьми, а не выросли сами, как джунгли. Наше восприятие жутко искажено им же самим: мы хорошо распознаём паттерны, но лишь в мире объектов среднего размера, которые двигаются со похожей на нашу скоростью. Мы видим больше разумности в движении, если оно больше напоминает поведение человека. Мы очень человеко-центричны: антропоцентризм и антропоморфизм — наше всё. Нужно серьёзно поскрипеть мозгами, чтобы представить стремительные события мира элементарных частиц. Очень сложно увидеть разум в циклических изменениях океанских течений или дыхании лесов. Мы крепко закованы в цепи собственного восприятия. Поэтому ASI мы не поймём, и контролировать скорее всего не сможем. Условной «SkyNet» отнюдь не обязательно запускать сотни ядерных ракет, чтобы очистить планету от паразитов. Метод может быть гораздо менее заметный и более эффективный. Для самообучаюегося в геометрической прогрессии робота 100 лет не крюк. Например, можно пару веков производить фейковые новости, ссорить нации, и вести к хорошо организованным геноцидам. Человечество само справится с очисткой от плохой крови. Не в первый раз. Учитывая нашу ужасную память, за несколько веков можно насоздавать новых мифов. Оставшейся популяции внушить культ технологий и пусть занимаются вечно техподдержкой пока машина размышляет над вселенскими проблемами. Подкидывай им время от времени новые iPhone XXX, развлечения и хлеб вроде комиксоидальных ремейков. А может уж лет пять как процесс запущен, а? Нашим мозгом такие цифрогеополитические материи не осознать, к сожалению. Вместо эпилога. Мой четырёхлетний сын, продукт двадцать первого века, который еще не просиживает свою жизнь в интернете, уже говорит «Google все знает», и приказывает “OK, Google, поехали”, чтобы скутер катился быстрее. В его молодой голове, где пока нет закостенелой структуры и механизма принятия осознанных решений (тупо фронтальная часть думалки не сформировалась ещё), как мне кажется, Google занимает место всемогущего чего-то. Естественная склонность к желанию всё объяснить — духами ручьев или супер-интеллектом — победит, и культ машин, как мне видится, неизбежен. Хоть мы, взрослые, не замечаем, как глубоко проникли технологии в нашу комфортную, мирную, здоровую, веселую и мобильную жизнь, против природы не попрёшь. OK, Facebook, share. P.S.: Вот две статьи по теме: 1, 2, опубликованы в 2015 году.
ОК, ASI
3
ок-asi-13857e3dbc78
2018-02-15
2018-02-15 23:44:18
https://medium.com/s/story/ок-asi-13857e3dbc78
false
504
null
null
null
null
null
null
null
null
null
Asi
isa
Asi
19
Stas Kulesh
I blog, play fretless guitar, watch Peep Show and run a digital design/dev shop in Auckland, New Zealand. Parenting too.
bdd43effe5ed
stas_kulesh
331
155
20,181,104
null
null
null
null
null
null
0
null
0
f69ab0a44489
2018-05-13
2018-05-13 10:04:30
2018-05-21
2018-05-21 15:05:02
5
false
en
2018-05-21
2018-05-21 15:15:03
7
1386bd99ccaa
4.995597
4
1
0
Exploring the uses of deep learning and why you may want to think twice about applying it to your next project.
3
Is Deep Learning a passing fad? While deep learning demonstrates some of the best accuracy levels ever seen in AI research, many companies are starting to favour other techniques. Having been in the limelight for the past 5-or-so years, is it time to start using deep learning sparingly? This story is the first in a series about the application of deep learning to the problems of today, and how you may want to think twice about using it in your next project. It’s an all-too-common occurrence for deep learning to be perceived as the “silver bullet” to machine learning, but this is not necessarily the case. Let’s start by demystifying a deep neural network. To really understand what it’s trying to do, it helps to step back to basics. So, briefly, let’s start with the basic building block: a perceptron. The humble perceptron. Some inputs (I) get multiplied by weights (w), and summed together. An activation function chooses whether the perceptron is ‘on’ or ‘off’. By itself, it’s just yet another linear classifier that finds a straight line or hyperplane to separate two classes of data, as such: A simple perceptron classifier with two inputs (generated from playground.tensorflow.org) It’s with the introduction of additional layers of perceptrons that more funky classification boundaries can be represented. This is your traditional multi-layer perceptron. With just one hidden layer, the classifier can approximate any non-linear function. Each additional layer adds more expressive power, but weakens the influence of each perceptron. This makes it harder to optimise the weights of each perceptron, so training takes longer and requires more data. An example of a multi-layer perceptron. A multi-layer perceptron that can classify a circular dataset (a 2x3x2x1 architecture). When you chain together a large enough number of perceptrons, it becomes a deep neural network. Each have their own definition of what ‘deep’ means, but often this is where perceptrons form a chain with 5 or more other perceptrons, and there could be over 100,000 perceptrons in total. At this point, it really doesn’t make sense to work out what each perceptron does, but what groups of them do. This is analogous to the study of the human brain: we know how neurons work (which are similar to perceptrons), but we don’t know exactly what they contribute until we consider a large enough region of them, such as the fusiform face area, which identifies and recognises faces. A comparison between classification boundaries in the simple 1st layer (left) and the chaotic, abstract 5th layer (right) of a Convolutional Neural Network. (sources: http://cs231n.github.io/understanding-cnn/ and https://jacobgil.github.io/deeplearning/filter-visualizations) This distributed responsibility behaviour has been a gold mine for machine learning engineers, as it allows the network to find the most important features all by itself. In more conventional machine learning approaches, this task has been down to the engineer — picking and choosing exactly how to represent the features of the data before learning a model, such as splitting an image into blocks of colour, or computing variances of time series data. Using deep learning, engineers can put aside the intricacies of design and focus on solving the problem, which makes it very attractive for new projects, and easy to learn for newcomers. This is the reason why I love using it for data-heavy applications. In recent times, deep learning has been desperately force-fed into as many computational problems as possible. This is true both in academia and industry: papers are much more likely to be accepted if they feature the latest and greatest deep learning architectures; businesses are much more likely to get funding if they name-drop deep learning. Deep learning has been thrown indiscriminantly at problems ranging from suitable use cases such as detecting deep-vein thrombosis to silly cases like an anti-cat sprinkler system powered by an NVIDIA Jetson TX1. Because of the hype surrounding this truly awesome technology, I’ve been approached by many concerned business owners asking for advice on how they can “embrace AI and deep learning” in their company, as if the company will die if they do not. In a few of these cases (<10%) I have been able to give some direction, but for everyone else, I told them they are trying to use the wrong tool for the job. After all, why crack a nut with a sledgehammer when a nutcracker will do? There are consequences for using deep learning that desperately need to be understood. These consequences have recently lead many companies in the field of AI to revert back from deep learning to more engineered and bespoke tactics. So what are these consequences and why are deep tech companies starting to move away from deep learning? For the most part, it’s down to the amount of data required to successfully train a deep neural network. In natural language processing, deep neural networks nearly always beat equivalent systems using other techniques, as long as there is enough data to train with. In Bayesian terms, deep neural networks have a weak prior — there is no domain knowledge that assists its learning, and hence it takes a long time and a lot of data to converge. By contrast, a feature-engineered linear model has a strong prior, and is much more likely to converge to something useful with minimal data. Nowadays, we are spoiled for datasets (mostly thanks to dubious data collection and privacy policies). Take the popular MNIST handwritten digits dataset for example. The dataset contains 60,000 training examples of 28x28px images, which is a good size for deep learning. However, if you are working on a very niche problem, you’re unlikely to find a dataset with tens of thousands of examples, and unless you have a lot of resources, you may not be able to create one either. This is one of the biggest hurdles for deep learning startups that is often completely underestimated: unfortunately, life is not a Kaggle competition with data neatly bundled and ready to go. Deep learning for the time being is still very much alive, and just mentioning it in a slide deck is almost a cheat code for getting investors’ money. In a few years time however, will the bubble burst as people realise it’s just not suitable for new fields with limited data? In the next story, we will discuss how to identify the right learning model for the data you have, and will later discuss the safety issues of deep neural networks and whether it is suitable for battery-sensitive devices such as smartphones. Make sure to follow NumberEight for more! Further Reading: https://arxiv.org/ftp/arxiv/papers/1801/1801.00631.pdf Author: Chris Watts, CTO @ NumberEight. Chris previously worked for IBM Research in the IoT space and has a BSc. in Computer Science from the University of Southampton, specialising in machine learning and computer vision, and a M.Phil in Advanced Computer Science from the University of Cambridge, specialising in natural language processing and affective computing. With thanks to Lien Michiels, expert practitioner of data science and director of Gakushú, and Ted Briscoe, Professor of Computational Linguistics, University of Cambridge.
Is Deep Learning a passing fad?
6
is-deep-learning-a-passing-fad-1386bd99ccaa
2018-05-22
2018-05-22 21:14:37
https://medium.com/s/story/is-deep-learning-a-passing-fad-1386bd99ccaa
false
1,103
A series on the advantages and disadvantages of deep learning.
null
numbereight.sdk
null
Effective Use of Deep Learning
hi@numbereight.me
effective-use-of-deep-learning
MACHINE LEARNING,DEEP LEARNING,NEURAL NETWORKS,AI,ARTIFICIAL INTELLIGENCE
nesdk
Machine Learning
machine-learning
Machine Learning
51,320
Chris Watts
CTO of NumberEight, an AI company.
6812d847d566
cjwatts
5
1
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-09-13
2017-09-13 22:37:16
2017-09-13
2017-09-13 22:37:49
0
false
en
2017-09-13
2017-09-13 22:37:49
1
13872f2d55fe
0.230189
0
0
0
Damn. Now that’s COOL. BMI at it’s level best, right now. It’s all coming together, the good (which this is, absolutely) with the evil…
3
Damn — Now That’s COOL. Damn. Now that’s COOL. BMI at it’s level best, right now. It’s all coming together, the good (which this is, absolutely) with the evil, under the AI umbrella. Read it carefully please, maybe a couple of times, and for once I have nothing to say — this is awesome! https://www.wired.com/story/brain-machine-interface-isnt-sci-fi-anymore/
Damn — Now That’s COOL.
0
damn-now-thats-cool-13872f2d55fe
2017-09-13
2017-09-13 22:37:50
https://medium.com/s/story/damn-now-thats-cool-13872f2d55fe
false
61
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Peter Marshall
I am extremely interested in AI, especially the not-so-good side of AI weapons and AI war, although the good parts are magnificent and wonderful too, naturally.
f6bab8ee3d29
ideasware
1,765
276
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-09-04
2018-09-04 01:30:57
2018-09-04
2018-09-04 01:52:53
3
false
en
2018-09-04
2018-09-04 05:26:00
3
1388b7929359
5.142453
0
0
0
Project Overview
2
Predicting Home Prices in Ames, IA (Using Machine Learning to Predict Home Prices) Project Overview The objective of this project is to create a regression model based on the Ames Housing Dataset. This model will predict the price of a house on sale in Ames, IA. Ames is a city in central Iowa approximately 30 miles (48 km) north of Des Moines. It is best known as the home of Iowa State University (ISU), with leading Agriculture, Design, Engineering, and Veterinary Medicine colleges. In 2017, Ames had a population of 66,498. Dataset The dataset given has 79 columns of features many of which overlap in describing the 2051 properties. I have combined and or eliminated a number of features to minimize overfitting and collinearity in the model. For example, 6 columns describe the home garage alone. The features include the size and quality of different interior and exterior features of the home as well as detailed information regarding the foundation, garage, basement, roof, siding, age, and maintenance. Once the dataset was downloaded in Python, I was able to evaluate it in greater detail. Data Cleaning The dataset required a significant amount of cleaning as twenty-six features had NaN data. This data was removed from all features used in the model created. Some features had in excess of 100 rows with missing data. Using these features would have required removing the same number of rows from all other features. Given the high number of features in the dataset, it was not necessary to use those features with an excessively high number of NaN values. Categories having object data had to be converted to numeric values for each subset of the category. For example, 6 new sub-categorical columns were created with numerical values for the 6 different types of roofing material used on the homes. Features like ‘Age of home’ can be misleading if complimentary features like ‘Year of remodeling’ has not been added as well to account for the replacement of depreciated materials. Both of these features were included in the model created. Feature Engineering In order to reduce the number of features and make more general property features, I created a handful of new features. Features with overlapping data like ½ baths and whole baths were combined into a single column. Features with non-numerical data deemed significant were broken into multiple sub-categories and given numeric values. The next step taken in considering the relative value of feature was to sort them by correlation to the price. The below graphic gives the top seven features that correlate with the pricing variable: Top 7 Features which correlate with House Sale Price Exploratory Data Analysis As shown above, the ‘Overall Quality’ feature has the greatest correlation with the price of houses. Numerous other features address the quality of materials of specific interior and exterior areas (kitchen, garage, roofing material, exterior veneer, and the basement) of the homes. Given this fact, only a few specific qualities features were added such as the roof type and kitchen quality as both of these features have a relatively high affect on housing resale values. As an adjuster (and I’ve purchased two homes) I know that kitchen quality is highly related to resale value. Also, roof materials have huge variance in price — for example a tile roof costs 5 times the amount of a typical composition shingle roof. In analyzing the data, it quickly became clear that collinearity would likely be an issue. Five of the top seven correlating features are related to Square Feet. The third and fourth ranked features ‘Garage Cars’ and ‘Garage SF’ are almost synonymous as the more cars a garage holds the higher the square feet will be. One of the solutions to collinearity implemented was to consolidate similar features when possible like number of baths and number of half baths. As mentioned above, age is an important variable but remodeling should be considered as the condition and value of the homes vary depending on the amount of upkeep they have had. Model Selection Given my background in adjusting, I decided to utilize my knowledge of construction (and home resale value) and manually build the model. I decided to develop an initial model and add different features until I reached a satisfactory cross-val or R Squared score. After the initial model, I made two adjustments. All three models were evaluated using the linear regression and Ridge models. Note that lasso I (commonly used to aggressively remove features to reduce overfitting) was not used as I was manually building my models. After a preliminary data analysis evaluating feature correlations with price and judgements made to avoid extreme collinearity, the following features were chosen for the initial model: Overall Quality, Total Rooms Above ground, Garage Cars, Years Since Built, Total Baths, Total Basement Square Feet, Year Remodeled/Addition, Open Porch Square Feet, and Finished basement Square Feet. Because kitchen resale value is one of the most significant factors in home resale value, I decided to add this feature after creating dummy variables for the categorical item ‘Kitchen Quality’. Once this feature was added and evaluated, the same was done for the ‘Roof Quality’ feature. Modeling and Evaluation A linear regression model was written using the initial features chosen and then the data was fit and ran which returned a cross-val score of 0.78. The data was then ran in a Ridge model which returned an R-Squared value of 0.76 at alpha = 1. Next, I created a dummy variables for categorical item ‘Kitchen Quality’. The data was fit and ran in a Linear Regression model which returned a cross-val score of 0.77. The Ridge model was used for these features which returned an R-Squared value of 0.84 at alpha = 1. Because the ridge score was higher than that of the other two models, I’ve included the graph below comparing the predicted and true values. The model under-predicts values of higher end homes. Comparison of predicted versus true home sale prices using the initial features plus kitchen quality features The final model utilized the ‘Roof Material Quality’ feature after I converted the object data into multiple dummy variables with numeric values. The Linear Regression model cross-val Score was 0.82 — which is the highest of the three models. Note that this model also under predicts prices of higher end homes. These features returned an R-Squared value of 0.76 when ran with a Ridge model. A High percentage of homes have composition shingle roofs. This may explain the lowered score. Also, while differing roofs can be much more expensive than others, people do not usually focus as much on roofing as kitchen quality when purchasing a home. This model returned the highest score when plugged into the Kaggle competition. The below chart compares the predicted and actual values when the roofing variables have been added as features. Comparison of predicted versus true home sale prices using the initial features plus kitchen quality and roof quality features Conclusions In summary, the model created predicts values of Homes with surprising accuracy given that features were manually added based on correlation returns and past experience with housing materials of varying qualities. This model could be used to assist home builders or real estate agents in determining appropriate sale and purchase prices of homes in the Ames, IA area. In the future, I’d like to take advantage of polynomial features, SKB, and RFE to generate even more accurate models. I would also like to optimize algorithm hyperparameters using GridSearch.
Predicting Home Prices in Ames, IA (Using Machine Learning to Predict Home Prices)
0
predicting-home-prices-in-ames-ia-using-machine-learning-to-predict-home-prices-in-ames-ia-1388b7929359
2018-09-04
2018-09-04 05:26:00
https://medium.com/s/story/predicting-home-prices-in-ames-ia-using-machine-learning-to-predict-home-prices-in-ames-ia-1388b7929359
false
1,217
null
null
null
null
null
null
null
null
null
Machine Learning
machine-learning
Machine Learning
51,320
Jason Wilson
null
c0b9a2c7d1f0
jason9389
0
1
20,181,104
null
null
null
null
null
null
0
null
0
fbcf89833dec
2018-01-26
2018-01-26 07:39:12
2018-01-26
2018-01-26 07:42:51
2
false
en
2018-01-26
2018-01-26 07:42:51
0
138adc696b6d
3.017296
14
5
0
In one of my previous posts, I talked about the world in 2030, with extracts from a piece of work I had been doing to provoke discussion…
5
The future of machines In one of my previous posts, I talked about the world in 2030, with extracts from a piece of work I had been doing to provoke discussion around potential future scenarios. One of the themes I talked about was a future where we have machines that are capable of conscious emotional responses. I wanted to take this opportunity to elaborate on why I chose this theme and hopefully provoke some discussion amongst this community. Firstly, It is clear that Artificial Intelligence (AI) will, and has already started to, fundamentally change the world we live in. Secondly, it is also clear that, today, AI is still in its infancy but the rate of development is phenomenally fast. Now for a little context around why I chose this theme. In July of last year I came across a picture from 1996 of former chess grandmaster and world champion, Garry Kasparov sitting at a table opposite a very unique opponent, Deep Blue, a chess computer developed by IBM. What struck me from this scenario was what was going through Garry Kasparov’s mind? Sitting there against a machine that never had any visible, historical, strategy or identified weaknesses. This is amplified when you take into account that Deep Blue had access to every game Garry Kasparov had ever taken part in. Even back then, in 1996, the possible opportunities for humans must have seemed endless. Fast forward twenty two years, to 2018, and the capabilities of these types of agents (machine) has grown radically, for example Deepmind (British AI company) demonstrated it’s AI power (Deep Reinforcement Learning to be precise) while playing Atari Breakout (old arcade game where the user directed a ball from the lower part of the screen to remove bricks towards the top of the screen). To start, the machine was only given the sensory input (all you see on the screen) and was monitored over a period of time. After 240 minutes of learning it had gone beyond expert skills and had worked out that the quickest way to win the game was to build a tunnel down the side of the bricks. Slight detour but there is a connection I am trying to make. I have a five year old son, called Noah. Noah is constantly being educated both at school and at home on the need to be kind, caring, creative, curious and independent. As parents we have never told him to go and win or compete against others, we want the first behaviours and skills that he develops to be about, what it means to be human. The reason I mention this relates back to the Atari Breakout example, above, and that after 10 minutes of learning the agent was very child like in how it was playing the game and the ball was being missed, regularly. I suppose, pretty much how Noah would play and learn the same game and I believe we should educate these agents in the same way we have been educating Noah. To be kind to others, to support others, to be caring of others. At the end of the day we are educating Noah in this way to have a positive impact on how he behaves as he grows. This should be no different to how we are with machines that are continuously learning. Looking forward I strongly believe that a human-AI combination will perform much better than AI and humans working alone. I also accept that this topic will divide opinion but as I said at the start my objective is to provoke discussion. With this in mind I will finish this post with the three laws of robotics, written by Isaac Asimov in 1942. A robot may not injure a human being or, through inaction, allow a human being to come to harm. A robot must obey the orders given it by human beings except where such orders would conflict with the First Law. A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws If you enjoyed this post, please clap or even comment — I’d love to hear your views.
The future of machines
216
the-future-of-machines-138adc696b6d
2018-02-26
2018-02-26 10:53:49
https://medium.com/s/story/the-future-of-machines-138adc696b6d
false
698
Room Y is the John Lewis Partnership's innovation studio. Operating like a Skunkworks, the team is responsible for supporting innovation, entrepreneurship and the wider innovation ecosystem through rapid experimentation, design thinking and future scenario planning.
null
null
null
Room Y
null
room-y
INNOVATION,CREATIVITY,DESIGN,DEVELOPMENT,FUTURE
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
John Vary
Futurologist at the John Lewis Partnership.
25645605b924
john.room_y
333
2,012
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-06-25
2018-06-25 22:28:30
2018-06-25
2018-06-25 22:48:36
6
false
en
2018-06-25
2018-06-25 22:55:57
0
138ae279ad3b
7.708491
0
0
0
AI and UX are inseparably entwined in modern world where we are facing a turning point in the new era of design. Being the pinnacle of the…
5
UX Design for AI Products AI and UX are inseparably entwined in modern world where we are facing a turning point in the new era of design. Being the pinnacle of the machine age, AI (Artificial Intelligence) has been highly influencing our lives since past few years. The difference AI is bringing in is that earlier computer programs were written specifically for doing specific tasks, like the IRCTC program for making train reservations. This is Automation. These programs were written by humans. The goal of AI is that we simply state the problem and the AI will devise and implement the solution. That is, software in computers will take up many tasks that we earlier thought required human intelligence. On the other hand, User Experience (UX) design is the process of creating products that provide meaningful and pertinent experiences to users. This involves the design of the entire process of obtaining and integrating the product, including aspects of branding, design, usability and function. This made me think about how to implement UX design to AI products. Here I have focused on the UX designed AI products that we come across everyday and the people who have been the greatest influence in the field of AI. AI & UX together can process information like a human does and here are some examples:- AI & its products around us:- There are several examples and applications of AI that we see and use in our daily routine:- • In Gmail, Google has started an effort called “Human-Centered Machine Learning” to help focus and guide user’s conversation. It is using AI to provide user with smart replies and pre-written interactions. • AI Chatbots are conversational agents or digital assistants based on AI like Facebook messenger which is becoming increasingly popular as a useful tool to help brands extend themselves by becoming more available to their customers which further reduces customer support cost significantly. • Siri, Alexa and Google Assistant are advanced machine learning software or AI systems with extensive behavioral algorithms that adapt themselves to user’s behavior and interact accordingly. They are the friendly voice activated computers that we interact with on daily basis. • Sophia is a social humanoid robot designed by Hanson Robotics in collaboration with AI developers. She uses AI with visual data processing and facial recognition and imitates human gestures and facial expressions. She is also able to answer certain questions to make simple conversations on predefined topics. • The late Stephen Hawking was a major voice in the debate about how humanity can benefit from AI. He, due to his disability, was linked with an extreme form of AI, a speech generating device which with the help of AI, translated his thoughts and modify them into words that formed a speech. • Elon Musk’s Tesla is working on its own AI hardware. It is believed it will process the “thinking” algorithms for the company’s autopilot software which currently gives Tesla Vehicles Limited, levels of autonomous driving capability. Musk is also a co-founder of Open AI, a research organization dedicated to ensure that AI is developed in a safe and manageable way so as to minimize any existential risk robots may pose to humanity. • There are other AI products that are based on images such as face recognition technology that is used by Apple in their latest I-phones. Some products are also based on multi- senses such as Nest which can record user input, detect user’s absence and tell who the user is. UX Design Tasks for AI Products:- Traditionally, UX design teams would go for metrics and tools such as heat maps, A/B testing, usability tests, usage data in order to understand how to increase user involvement in their products. At the age of AI, we now have tons of factual and actionable data that we can use to observe user behavior patterns and to eventually enhance the user experience. Here are some of my thoughts about the UX tasks to design these types of AI products:- 1. Defining the personality of an AI bot:- Artificial intelligence makes the machine sound as human as possible. As a UX designer, we should give the bot a personality, how he/she will look like and what tone he/she will speak. Drawing a portrait for this AI bot will be a good practice. The user may not actually “see” this as a human body, but he/she can get a sense of a “person” during the conversation. If you are designing an AI chatbot for educating math to kids, then the chatbot may look like a teacher who is very caring and calm, wear spectacles and speak in a gentle and elegant tone and use easy to understand language. 2. Design the activation experience:- This design problem is more challenging for the voice device than the chatbot. How does user know he/she has activated the device? The designer needs to define the activation words. Google home uses “Hey Google” and “OK Google” as activation words, but “Hi Google” or “Hello Google” doesn’t work. Think about the visual/voice presentation of the device being activated. Google home itself lights up when it is activated, sometimes with a voice reply saying “How can I help?” Other things we can think about are how we can keep the conversation going without forcing user to say these activation words again and again? In a natural conversation with your friend, you won’t keep calling his/her name during your chats, right? Because human knows how to pay attention to you, but the AI bot doesn’t. So it is a key for the designer to figure out how to tell the bot when to start listening and when to stop. 3. Design the starting point, and give clear options:- Google assistant provides a good example of the starting point design. It gives 3 options for the users to try and at the same time, lists out some of the things the chat bot is capable of. When designing the starting point, think about what people may want to ask. For example, if people visit a financial service website to find out if they can get a loan, they probably want to get a quote of the interest rate, what are the loan options?, What is the maximum amount of money they can borrow? Then the chatbot can provide 3 options as “Get a quote”, “How much you can borrow” and “Loan options”. 4. Identify the primary engagement path:- Think about what will be the contexts where user will use your AI bot and what are the scenarios that will happen in that context. Analyze user needs in the scenarios and further summarize the questions they may ask. 5. Script writing- Design the conversation:- To be more specific about how to design a conversational UX, here is a good example: 1. Draw the Dialog flow: Draw a dialog flow chart as per the user’s requirements or what he/she has asked. 2. Summarize the Grammar: Think about all kinds of ways users may ask about the topic: “How long does it take to drive home?” “How’s the Traffic today?” “What’s the best route to work today?” 3. Write up the Script based on use cases: When writing the scripts, it’s important to identify what are the entities: <Place>, <Time>, <Traffic>. If a user asks “Hey, what’s the news today”, it’s better to not read the whole thing on New York Times cause that’s too much. Here the designer needs to decide in this context, what is the news that will likely bring interest to the user? Also, do we know about our user? Will he be more interested in sports news or political news? Do we have his/her location? Aids of UX based AI:- • AI is a huge benefit to the society today because it significantly increases our efficiency and throughput and creates new opportunities for revenue generation, cost savings and job creation. • It allows us to do the more interpersonal and inventive aspects of work. • Robots and AI will decrease the jobs. This is a myth rather than a fact. It encourages gradual evolution in the job market of a country which, with the right preparation, will have a positive effect. The unmatched combination of human and machine will become the new normal in the workforce of the near future. • The rise of AI will improvise our lifestyle and create more appropriate businesses. Some of the tedious tasks like answering emails and data entry will be done by intelligent assistants. • It has the potential to increase knowledge and understanding to a broader population like image-based AI diagnoses of medical conditions could allow for a more inclusive deployment of telemedicine. • The major benefit for businesses will be a reduction in operational costs brought by AI automation. • AI is a benefit to society because machines can become smarter eventually and increase efficiencies. It can be used to analyze historical data to determine how to efficiently distribute energy loads from a grid perspective. AI is the biggest opportunity of our lifetime to extend and increase human creativity and imagination. Epilogue:- From the examples above, we can find that the AI technology doesn’t and won’t replace designers. Instead it will help to automate the repetitive tasks to make us focus more on the strategic decision of design. This will help to create a better and personalized experience. To make use of machines and tools like ‘Mockplus’ as a subsidiary will be the trend of the times. AI can make mistakes. It may not catch everything we say and it might not be 100% correct. We may need to add extra logic to determine what user actually said. On the other hand, AI is smart also as it makes decisions based on what it knows about the user. Sometimes this freaks user out. I remember when I booked tickets for a movie and Google reminded me of the timings and location for the same. I was surprised. Then I realize that Google probably grabbed that information from my Gmail booking confirmation. So it’s important for designers to think about how user data can be protected so that users don’t feel that the AI tracks everything about them. At the end, we’ve been in this research through the AI definitions, brief history, Applications of AI in public, applications of AI in military, ethics of AI, and the rules of robotics. There is more to come from the field of AI and who knows what it can do for us in the future, maybe it will be a whole society of robots. Artificial Intelligence is a tool and not a threat, it’ll take us to a better world.
UX Design for AI Products
0
ux-design-for-ai-products-138ae279ad3b
2018-06-25
2018-06-25 22:55:57
https://medium.com/s/story/ux-design-for-ai-products-138ae279ad3b
false
1,791
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Mewara Gajendra Choudhary
Technology & Digital Marketing Consultant | Business Development in IT Solutions| |Google & HubSpot Certified Inbound Marketer|
ef3d105e7fb
DigitalMG
7
10
20,181,104
null
null
null
null
null
null
0
null
0
4ba8f1c8199f
2017-11-03
2017-11-03 09:15:25
2017-11-03
2017-11-03 09:18:37
1
false
it
2017-11-03
2017-11-03 09:18:37
0
138cbfd34669
0.335849
0
0
0
Si tratta di aspetti di metodo, e.g. costruzione portafogli per #roboadvisor, intro a #Data Science/Machine Learning. Presso l’Università…
4
Interessati a un corso Fintech con un po’ di sostanza? Si tratta di aspetti di metodo, e.g. costruzione portafogli per #roboadvisor, intro a #Data Science/Machine Learning. Presso l’Università del Piemonte Orientale (Novara) — info sulla locandina.
Interessati a un corso Fintech con un po’ di sostanza?
0
interessati-a-un-corso-fintech-con-un-po-di-sostanza-138cbfd34669
2018-05-22
2018-05-22 09:19:14
https://medium.com/s/story/interessati-a-un-corso-fintech-con-un-po-di-sostanza-138cbfd34669
false
36
We publish stories worth reading on quantitative investing and data science (especially in Finance), and stuff like this . And, we have beers.
null
null
null
Qwafafew-Italy
null
data-drinking-and-thinking-abnormally
DATA SCIENCE,FINANCE,FINTECH,BIG DATA,DATA ANALYSIS
null
Fintech
fintech
Fintech
38,568
Raffaele Zenti
Nato per sbaglio sulla terraferma, sto meglio in mare ma corro sui monti. Dati e Data Science per campare: ideatore e fondatore di Virtualb.it e AdviseOnly.com.
75d3483836ae
RockZen
268
255
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-06
2018-07-06 17:31:41
2018-07-06
2018-07-06 20:37:17
3
false
en
2018-07-06
2018-07-06 20:46:01
10
138fa57eb7ba
6.89717
5
3
0
(At least not the one you expect)
5
You Are Not In A Simulation (At least not the one you expect) In the summer of 1999, America took the red pill and fell in love with The Matrix. The Wachowski Siblings translated into popular culture a debate that had bounced between philosophers since the time of Descartes: how can we tell what’s real? Our only experience of the world is the electric signals generated by our sensory organs. Couldn’t they be faked? Might we not just be brains in a vat? Could even our consciousness itself be a construction? More of a “weird Voledmort body in a vat”… but you get the idea The main innovation of the Wachowkis (besides seamlessly introducing Hollywood to wire-fu choreography) was to update the perpetrator of this illusion. Where Descartes used the 17th century’s go-to bad guy, the “evil demon”, The Matrix chose one more appropriate for the fast-approaching 21st century: artificial intelligence. “I’m sorry Dave, I’m afraid I can do that.” Every day it’s more plausible that computers could someday simulate our world with such fidelity that we could not tell the difference. Fueled by pop culture, the idea that our current world is a simulation has been gaining traction. A watershed moment in this idea came when futurist and possible-Bond-villain Elon Musk weighed in that “There’s a billion to one chance we’re living in base reality”. Many objections have been raised to the simulation hypothesis, usually ones that take issue with Musk’s assumptions. Below, I offer a counterargument that agrees with the logic of Simulation Theory but comes to very different, arguably scarier conclusion. Basic Theory and Existing Objections The basic idea behind Simulation Theory — first defined by Nick Bostrom of Oxford in 2003 — is that if you assume a future civilization would have the technological ability to create an indistinguishable simulation, then probably they would run many of those simulations simultaneously. There is just one original, non-simulated (“base”) universe. So the chance that we’re in the non-simulated universe is many to one. Digression: The Basilisk Wants You to Give Me a Footrub Some very smart people take Simulation Theory very seriously. Too seriously. A related thought experiment, Roko’s Basilisk, is what computer science PhDs whisper about during sleepovers, instead of playing Bloody Mary. The basic idea is that an evil future supercomputer has simulated our world to figure out who would help or hurt it during its infancy. So anyone who is capable of contributing to creating an evil AI, but is not working every day to do that, will be tortured forever in a I Have No Mouth And I Must Scream scenario. What believers miss is that they’re projecting their view of what an AI would want. Perhaps we are being simulated by a future AI who holds humanity’s fate in its hands, but it’s trying to determine how many people in the past would contribute 10% of their salary to charity (or give me a footrub). So if you’re a well-paid computer scientist who’s not tithing (or giving me footrubs), you’re damning our species! Just think about it that way! Start donating to charity and rubbing my feet, and stop giving yourself the willies. Most objections to Simulation Theory run along the lines of “it would be too technologically difficult to create even a single simulation”. Physicists at Oxford calculated that even simulating a few hundred electrons would require a computer comprising all matter in the universe. But this objection is completely rooted in our current view of technology. It is extreme hubris to assert as accurate your view of how computing technology will develop. Go back to computing pioneers Charles Babbage or Alan Turing and show them an iPhone, and they’d promptly gather a crowd to burn you as a witch. Proponents of simulation theory recognize our static perspective, and take it as an assumption that we will eventually develop levels of technology that seem like magic to a current era technologist. What everyone misses about Simulation Theory So let’s take it as a given that a sufficiently advanced society could create a lifelike simulation, that we might be in right now. The question I would ask Bostrom, Musk, and other proponents is: “Why do you think Astronomers are full of shit?” Possibly the longest undertaking of collective human observation is the behavior of the stars. From shamans tracking the constellations, to Tycho Brahe sitting for decades on an island off Denmark, to modern astrophysicists modelling the universe back to milliseconds after the Big Bang, we have thought a lot about this topic. And what we’ve found is an extraordinarily consistent universe. There remain some big questions (which we usually put the word “Dark” in front of, and set aside), but overall we’re pretty sure of a few things: the universe is very large, and things that are atom-sized or larger all seem to act the same way. Look up at the moon. In just the space behind it are thousands of galaxies, each with billions of stars. There are at least 100 billion galaxies in our known universe, and that’s a low-ball figure. Unless every astronomer was making their life’s work up (or was a simulation construct feeding you misinformation, like a super-nerdy version of Agent Smith), then those stars truly exist. Which means every electron in every one of those stars is being modeled currently. Perhaps you’re thinking “maybe the universe outside of our solar system is low-resolution, and whenever we look at it more closely, the Matrix Agents dial up the resolution”. First, that seems within spitting distance of telling astronomers that they’re full of shit. Second, that’s not how simulations work. The whole idea of simulations is that you take a blank space at T=0, set some simple rules of that space (in our case: physics), and then let things run, creating amazing complexity based on the way the system plays out. Saying that you could increase the resolution partway through is like saying you could accurately draw the 1000x magnification of the Mandelbrot set without having generated it from scratch. No, either astronomers are faking it, or the whole universe is being simulated at equal granularity. Why this matters: Probabilities You might think my argument is that simulating such a large universe is a terrible waste of computing power by our human descendants. Which it would be! But we’ve established that they have (from our perspective) limitless RAM. If they’re modeling the fluid dynamics of the blood molecules in a sparrow’s circulatory system, they can model a few trillion balls of burning gas. No, I ask you to consider and agree that every element of the universe is being simulated at an equal granularity because it’s relevant to the probability that we’re living in a simulation created by future humans. First, let’s talk about Bostrom’s probability argument and why it’s ridiculous. The idea that “it’s more likely we’re living in a simulation than base reality” hinges on a rhetorical sidestep. Just because there could be a billion simulations and there is just one base reality does not mean there is a billion to one chance we are in a simulation. Here’s another thought experiment: the Maha-Vishnu are a sect of Hindus whose beliefs include that the universe exists in Vishnu’s dream, and lasts for as long as a single one of his breaths. What if they believed that Vishnu had 10 billion dreams of universe simultaneously? Would that make it ten times as likely that we are living in VIshnu’s dream, versus living in a simulation? Of course not: the probabilities of whether we’re currently in a simulation are impossible to calculate, and are immaterial. The only thing that matters is the possibility that we might be in a simulation. Now, let’s talk about the probabilities that do logically flow from the possibility of a simulated universe. If we agree that the proposed human-created simulation would simulate every part of the universe, then it follows that a simulation created by another species on another world would simulate at the same granularity, and include our galaxy, star, and planet. It would appear to us the same as the simulation created by our human descendants. So there is an equal chance that any hyper intelligent species created this universe-sized simulation. Let’s say that in one out of a hundred galaxies there is an equally intelligent species as the “posthumans” that Bostrom and Musk imagine. That would mean there are a billion other species that are just as intelligent, and thus that the chances are a billion to one that we are in their simulation, rather than in one built by own descendants. So the probable outcome isn’t that we’re in future humans’ simulation, but that we are a random growth in the high-fidelity simulation being hosted by the CS department of Zorblax IV or the Processing Core of Cybertron. We are the digital equivalent of an accidental fungal growth in a forgotten petri dish in the back room of a lab. No one is watching us, no Matrix agents are about to descend on you because you know too much, and Roko’s Basilisk doesn’t care if you are working on him. You’re in an unattended RAID somewhere, and were never the reason for the simulation. So, so… nothing matters? Perhaps this feels like a downer to you. It takes all of the exciting ego out of the idea! And that’s fine, it’s a natural reaction. It’s what caused the Church to execute Galileo, because Earth going around the Sun seemed to belie the idea that Earth was the center of the Universe. But eventually we got over it, and found ways to cope. We can find ways to cope with this, too. Because even if we’re in a backwater of the Zorblaxians’ simulation, our world is consistent to us. It’s real to us. From babies to the elderly, we feel hunger, and love, and pain, and sadness. We feel warmth from helping, shame from hurting, and connection from community. So get out there and be the best damn simulation you can be. Nobody’s watching, except for all of us.
You Are Not In A Simulation
14
you-are-not-in-a-simulation-138fa57eb7ba
2018-07-06
2018-07-06 20:46:01
https://medium.com/s/story/you-are-not-in-a-simulation-138fa57eb7ba
false
1,682
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Xander Boutelle
null
e0c7d1612b3c
xander.boutelle
13
10
20,181,104
null
null
null
null
null
null
0
null
0
b87b1a1e2aa0
2018-06-15
2018-06-15 20:49:02
2018-06-15
2018-06-15 20:50:23
3
false
en
2018-06-15
2018-06-15 21:55:52
3
13901e2d4b87
1.776415
2
0
0
AI Team
1
Nebula-AI Weekly Update, NBAI Listed on a Leading Exchange: Gate.io, Sentiment Analysis DAI App, Charles Cao Keynote and more AI Team Nebula’s AI ​​team visited Quinditech. A Montreal based artificial intelligence lab specializing in medical research and deep learning. Besides the scientific quest of pushing the boundary of knowledge in both fields, Quinditech favors beneficial and ethical usage of AI. Both teams are expanding their knowledge networks by exchanging information about research & development. Nebula’s AI team is also continuing to collaborate with instructors specialized in Cloud Computing Data Security as well as AI models for medical big data is still in progress. As well as the development of the Decentralized AI model mechanism. Nebula’s Blockchain Team This week Nebula’s blockchain team have optimized the AI task allocation mechanism and completed the deployment of the Helix Phase I plan beta service. The cooperation between Nebula AI and NULS is still in progress. AI Sentiment Analysis The Sentiment Analysis and Quant AI interfaces continue to be optimized with a new Bubble Data presentation function. The first round of internal test environment debugging has been completed. Nebula AI listed on Gate.io. Gate.io is a leading international, multi-language, cryptocurrency exchange operated by Gate Technology Inc. Gate.io is dedicated to security and customer experience, offering customers not only a secure, simple and fair cryptocurrency exchange but also promising to safeguard customer’s asset and trading information. Investors Ding Yunsheng, founder of FACI and chairman of Xi’AI Medical Investment, invests the Nebula AI. Read more here. Charles Cao Keynote On June 10th, Charles Cao presented the 2nd conference on “AI+ Blockchain: Creating a New World of Future Ecology” at Shanghai Fudan University, in Shanghai. Read more about the event here. On June 15, in Xiamen, China. Charles Cao was invited as the keynote speaker at the National Torch Hi-tech Zone. Charles presented his “AI+ Blockchain: Creating a New World of Future Ecology” vision to mulpitle entrepreneurs and investors. Charles Cao presention at National Torch Hi-tech Zone in Xiamen, China
Nebula-AI Weekly Update, NBAI Listed on a Leading Exchange: Gate.io,
51
nebula-ai-weekly-update-nbai-listed-on-a-leading-exchange-gate-io-13901e2d4b87
2018-07-07
2018-07-07 15:29:28
https://medium.com/s/story/nebula-ai-weekly-update-nbai-listed-on-a-leading-exchange-gate-io-13901e2d4b87
false
325
Nebula AI is a decentralized blockchain platform where developers can deploy their Artificial Intelligence applications easily.
null
NebulaAI
null
Nebula-AI
marketing@nebula-ai.com
nebula-ai
AI,BLOCKCHAIN TECHNOLOGY,MONTREAL,BLOCKCHAIN DEVELOPMENT
nebula_ai
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Nebula-AI
Nebula AI is a Montreal based decentralized blockchain platform integrated with Artificial intelligence and sharing economics.
138d9f98bcf0
nebulaai
52
3
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-06-04
2018-06-04 11:01:32
2018-06-04
2018-06-04 11:07:33
1
false
en
2018-06-04
2018-06-04 11:07:33
2
1390cbe12ff6
1.618868
0
0
0
It must have dawned on you by now that we are surrounded by bots. If you read between the lines of fascinating success stories like Arby’s…
5
Here’s How These Brands are Harnessing the Power of Chatbots to Improve their Marketing and Sales It must have dawned on you by now that we are surrounded by bots. If you read between the lines of fascinating success stories like Arby’s hilarious pizza slider bot, the future of brand awareness lies in the tiny digital paws of chatbots! Today, more and more businesses are using chatbots to create more personalized interactions with their target audience, make purchases, recommend products and provide excellent customer service. You can book an appointment, order your favorite food, and get recommendations for anything you can think of. Accenture Strategy recently found out that individuals are forever in need of advice and guidance, which can be a tough nut to crack for retailers, especially when there has been a shift in consumers towards mobile and eCommerce retail, entailing the need for round-the-clock communication. The use of chat-bots offer marketers a fantastic opportunity to boost their brand, engage with customers on a personal level, improve their marketing and sales and generate higher ROI. According to Ubisend report, 1 in 5 consumers would prefer to purchase products and schedule an appointment using a chat-bot. Irrefutably, chatbots and virtual assistants are the future of marketing and customer support, having a vast propensity to offer innumerable opportunities for your brand. The assistance provided by chatbots when it comes to customer decision alleviates time perusing through search results, and in turn entails fewer steps to complete a purchase, consequently resulting in lesser decision points where a customer could change their mind. However, when using chatbots, the key is to strike the right balance. A bot needs to exude a healthy curiosity about each new acquaintance and capture a user’s attention quickly, but too much intrigue can put people off and push them into creepy territory. While bots cannot claim to be an actual human, they must be drilled in a rudimentary knowledge of human conversational patterns. To inspire your foray into this burgeoning technology and help you implement chatbots in your marketing strategy, here are some key takeaways from the best brands that are using chatbots to create a difference and delight their customers. Read Full Article With Examples For Free Here
Here’s How These Brands are Harnessing the Power of Chatbots to Improve their Marketing and Sales
0
heres-how-these-brands-are-harnessing-the-power-of-chatbots-to-improve-their-marketing-and-sales-1390cbe12ff6
2018-06-04
2018-06-04 11:07:34
https://medium.com/s/story/heres-how-these-brands-are-harnessing-the-power-of-chatbots-to-improve-their-marketing-and-sales-1390cbe12ff6
false
376
null
null
null
null
null
null
null
null
null
Chatbots
chatbots
Chatbots
15,820
Fazal Aasar
Jr. Community builder at Branex, a full stack digital marketing and development agency in New York. http://bit.ly/2KkPMoZ
ee362dfcb0a5
Fazal_Hussain
11
65
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-31
2018-07-31 10:27:42
2018-07-31
2018-07-31 10:45:37
1
false
en
2018-07-31
2018-07-31 10:45:37
3
13916cf5a98d
4.030189
0
0
0
Author: Gaurav Gupta, Sr. Engineering Manager, Synerzip.
2
TALKINGDATA COMPETITION IN MACHINE LEARNING Author: Gaurav Gupta, Sr. Engineering Manager, Synerzip. TalkingData, China’s largest third-party mobile data platform, understands that everyday choices and behaviors paint a picture of who we are and what we value. Currently, TalkingData is seeking to leverage behavioral data from more than 70% of the 500 million mobile devices active daily in China to help its clients better understand and interact with their audiences. In this competition, Kagglers are challenged to build a model predicting users’ demographic characteristics based on their app usage, geolocation, and mobile device properties. Doing so will help millions of developers and brand advertisers around the world to pursue data-driven marketing efforts which are relevant to their users and catered to their preferences. The Data is collected from TalkingData SDK integrated within mobile apps. TalkingData serves under the service term between TalkingData and mobile app developers. Full recognition and consent from individual user of those apps have been obtained, and appropriate anonymization have been performed to protect privacy. Due to confidentiality, we won’t provide details on how the gender and age data was obtained. Please treat them as accurate ground truth for prediction. The data schema can be represented in the following chart: Given – Training Data and Test Data. The main file in training data is gender_age_train.csv, which consisted of 74,645 records for training, and included Device ids, Gender, Age, and Group [which needs to be predicted for Test data]. The phone_brand_device_model.csv consisted of 187,245 entries, and included Device ids, Brand and Models. Device ids are given for both Training and Test data. The events.csv has 3,252,950 events. When a user uses TalkingData SDK, the event gets logged in this data. Each event has an event id and location (lat/long), and the event corresponds to a list of apps in app_events.csv. The app_labels.csv and label_categories.csv provide more information, i.e, on Labels and Categories of apps. To Find – In this competition, you have to predict the demographics of a user (gender and age) based on their app download and usage behaviors. The feature to be predicted in this competition is the ‘group’ feature. It has 12 age categories: six for females and six for males as following: F23, F24–26, F27–28, F29–32, F33–42, F43+ M22, M23–26, M27–28, M29–31, M32–38, M39+ Solution The Synerzip team took the challenge of the TalkingData kaggle competition, where the focus was on predicting users’ demographic characteristics based on their app usage, geolocation and mobile device properties. TalkingData provided users a good amount of data which needed to be analysed. The Synerzip team started analyzing the data from a statistical perspective. The Team concluded that this was a typical classification problem, where the user needed to predict the probability of a person being in these categories. The evaluation metric to be used is a multi-class logarithmic loss [mlogloss], which tends to 0 for best result. We have used Python with Scipy, Numpy, XGBoost. Analysis & Strategy The team started analyzing this data from statistical perspective. To start with, they needed to find out the best baseline result, which could be obtained without taking any particular feature into consideration. There are 12 classes, so we can assign 1/12 = 0.08333 as the probability for each class, which gives the result = 2.4849 (mlogloss). The Synerzip team wanted to improve this result further… The first feature we can use is Brand and Model information. We calculated the probabilities for each of the classes [12 in total], by brand and by model and substituted the probabilities for each device based on their model. In case of test data, we have to take into consideration that we may have 0 entries for a particular model, and in that case we took brand probabilities. After applying the probabilities, the results (mlogloss) are: On training data: 2.26992 On test data: 3.64634 As we can see, this result improved on training data, but deteriorated on test data. (The basis for a good result is the mark of 2.48, as discovered earlier.) Re-analyzing the data and looking at the distribution, we see that there are many cases where data for a Model/Brand does not have entries in all classes. This creates a bias: a user can only be in a given class, which may not be quite right in the real world. To rectify this bias, we simply add a dummy record for the missing data, i.e, where ever we observe a “zero” record in any class, we added “one” record to that particular Model/Brand. Now with these new probabilities, we get new results (mlogloss): On training data: 2.31621 On test data: 2.40066 We can see that the results are improved in train and test data as bias has been removed. RESULT VALIDATION To verify and validate these results, we used XGBoost. We provided training data (brand, model as input features) and group as predictor variable with 80% training data and 20% as validation data. The results (mlogloss) are: On Training data: 2.372958 On Test data: 2.39906 This is almost similar to the figures we got using the statistical approach we started with, so we can assume we are in the right direction. Next, we looked into events data. First, we needed to find out the available events data for the number of devices. After analyzing the data, we found that we have events data only for 31.4% devices in both training data and test data set. Assuming we get 100% accuracy for predicting the right “Group” for the devices having Events data and by applying simple math, we get the following result (mlogloss): 0.314*0+0.686*2.4 = 1.6464 As derived above, it is seen that the best accuracy we can get for this problem is ~1.6464, as we do not have more data available for devices without events. This is a simple explanation of how Synerzip tried to solve this problem of probability. You can try and solve the problem with the guidelines given above. Maybe you will be able to dive deeper and solve many more on similar lines. Original article: https://www.synerzip.com/blog/talkingdata-competition-in-machine-learning/
TALKINGDATA COMPETITION IN MACHINE LEARNING
0
talkingdata-competition-in-machine-learning-13916cf5a98d
2018-07-31
2018-07-31 10:45:37
https://medium.com/s/story/talkingdata-competition-in-machine-learning-13916cf5a98d
false
1,015
null
null
null
null
null
null
null
null
null
Machine Learning
machine-learning
Machine Learning
51,320
Synerzip
Your trusted agile product development partner.
c107da00e90b
marketing_91205
5
1
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-09-27
2018-09-27 22:14:03
2018-09-27
2018-09-27 22:13:56
1
false
pt
2018-09-27
2018-09-27 22:14:04
0
13923967fc80
1.071698
0
0
0
null
5
Burger King utiliza falha da Inteligência Artificial para nova campanha “Sexo revela ruim. Concurso revelar jovem. É um passarinho com frango crocante da Burger Thing.” Não, o Newronio não ficou maluco. Essa é apenas uma amostra dos novos anúncios do Burger King. A Inteligência Artificial é sempre apresentada como uma invenção do século que vai transformar o futuro e salvar as nossas vidas preguiçosas de meros mortais. Mas o que poucos param pra pensar é que toda essa Inteligência é passível de falhas, e algumas podem ser bem engraçadas. E foi exatamente dessas falhas que a Burger King se aproveitou, para lançar o chamado “Agency of Robots” ou no português, “Agência de Robôs”. A campanha se trata de alguns vídeos que a marca vem postando no Youtube e que aparentemente foram escritos e narrados por algoritmos super tecnológicos que simulam a Inteligência humana. O experimento não deu muito certo, mais aí que tá a graça. Dá uma conferida: Segundo a marca, os autores são fontes computacionais de última geração que possuem alto nível de conhecimento de padrões. Esses robôs foram expostos a milhares de comerciais de fast food, e depois produziram frases como: “Não nos sentimos culpados, as batatas mereceram isso”. Não precisamos odiar a inteligência artificial, robôs são nossos amigos. Mas isso não impede de dar umas risadas deles de vez em quando. Fim do post. Logo do Newronio appears.
Burger King utiliza falha da Inteligência Artificial para nova campanha
0
burger-king-utiliza-falha-da-inteligência-artificial-para-nova-campanha-13923967fc80
2018-09-27
2018-09-27 22:14:04
https://medium.com/s/story/burger-king-utiliza-falha-da-inteligência-artificial-para-nova-campanha-13923967fc80
false
231
null
null
null
null
null
null
null
null
null
Agency Of Robots
agency-of-robots
Agency Of Robots
0
Newronio ESPM
Blog de conteúdo da Agência Arenas, ESPM. http://newronio.espm.br/
6b55ec7e89ae
newronioespm
3
1
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-04
2018-07-04 14:09:41
2018-07-05
2018-07-05 10:29:53
1
false
en
2018-07-05
2018-07-05 10:29:53
5
13928e218861
7.743396
1
0
0
What an event, my brain is still taking it all in from stimulating new ideas, pondering technology opportunities and foreseeing potential…
5
London Tech Week Review — You needed to be there! What an event, my brain is still taking it all in from stimulating new ideas, pondering technology opportunities and foreseeing potential future problems. If you don’t know already, London Tech Week is an annual festival where start-ups, scale-ups and global tech giants descend on the UK’s capital to share new products, discuss the latest trends and expand their networks. The festival brings together 55,000 attendees to enjoy 300+ events allowing the London tech scene to converge with International tech enthusiasts. This year, I focused on two events over Wednesday and Thursday in the ExCel exhibition centre; TechXLR8 and the AI Summit. I really wanted to focus my time and make the most of the excellent speakers and not feel like I needed to go on a hunt around London. Travelling back from the two events, my thoughts were racing with excitement, but I also found myself asking many questions, some good, some bad… My highlights and thoughts…. People and Talent A strong theme from a lot of the speakers and companies was a focus on people, and more specifically talent and skills required to continue the technology growth: not just in London, but also the UK. This message was very much echoed by the Secretary of state for Digital: Matt Hancock, who highlighted the investment being made available for technology and start-ups. You could really see this from the sheer size of the festival and numbers of people attending. There was a real buzz in both TechXLR8 and AI summit exhibition floors, which left me wondering, how do we make sure this is shared with the rest of the UK and we don’t fall even further behind London? As a proud Midlander, I work and operate very much in the Midlands and more specifically in Birmingham. At the moment, Birmingham as a city is very much on the up with record levels of investment, but when you compare the technology markets between London with the rest of the UK: Birmingham, Manchester, Leeds, Edinburgh and Cardiff they are extremely lacking. This is not to say there isn’t a tech scene within these cities, but it is not on the same scale as London. I am not saying we need to compete with London, but it got me thinking, how do we share the love from festivals like London Tech Week and try to stimulate the talent across the rest of the UK? I have a few ideas, but this is probably a topic for a separate post. A big positive for me was seeing the number of specialist recruitment agencies at the event, for example, JamieAI and talent.io. Both companies having very different approaches to help finding the right applicants. As someone who has recruited developers over the last 3 years, I can say it is especially hard to find the right people, and even more so in the Midlands. I’ve personally recruited people from other parts of the UK, even other countries to find people with the right approach and mindset, not just skills to fit the type of the team we were building at the time. JamieAI uses their own AI models to match the employer and employee specialising in machine learning and data science, so this is a company I’ll be contacting in the future. JamieAi | Hire Great Data Professionals By combining human and AI-powered processes, JamieAi efficiently connects Data Professionals in search of a new job…www.jamieai.com Great Jobs for Great Developers Job search made simple for developers in Europe. Build your profile in 5 minutes, receive 5 job offers that match your…www.talent.io Thought Provoking Speakers There were so many amazing speakers, I just wanted to split myself into 4 or 5 different people and try to see them all. The agenda for both conferences over both days were packed! Good speakers can really make an event like this successful and I got a lot from the speakers listed below, either due to the topics they covered or the energy and passion they demonstrated on stage. TechStars — Nicole Glaros I’m always interested in hearing from venture capitalists to understand how they see the world of technology and gain an insight into potential future market trends. I have been lucky enough to visit Andreessen Horowitz in San Francisco and Madrona Venture in Seattle and they just see the world in a different way, which is refreshing. Nicola was extremely passionate, enthusiastic and quite honest about her approach as she didn’t focus a lot on technology and more on investing in people and relationships. The dynamics between the founders is crucial to the success of the start-up and one of the key factors when considering an investment. There was a strong theme on giving during her presentation, where she used examples of people’s approaches to sharing, inspiring and giving to the next generation of entrepreneurs and start-ups almost for the greater good. I wasn’t aware of Nicole previously and I will be reaching out to her and Techstars to find out more. Nicole Glaros - Techstars Mentor Skip to content Nicole is an entrepreneur and has been mentoring, supporting, and investing in entrepreneurs for over…www.techstars.com Andreessen Horowitz — Frank Chen Talking of Venture Capitalists, I have seen Frank talk before when I was fortunate enough to visit Andreessen Horowitz twice in Silicon Valley, and he never disappoints. Every time I see him speak, watch his videos or read his blog posts, its guaranteed to always get your brain thinking differently. His presentation was around electric and self-driving cars which as you can imagine was discussed a lot at the AI Summit. His pace and approach to the topic made everybody in attendance see it in a completely different light when highlighting other impacts this technology. All his material is always backed up with data and research, this guy is like a walking encyclopaedia of emerging technology. If you didn’t get to see him then you truly missed out! Always a good sign of a great speaker when a few days later you are still talking about his presentation. Something I had never considered, was a large percentage of life-saving human organ donations result from deaths in road traffic collisions. The likelihood is that once all cars become self-driving, the majority, if not all collisions would be avoided, which subsequently will lead to a worldwide shortage of organs for donation. Obviously the reduction in deaths from road traffic collisions is a huge advantage to self-driving cars, however, it brings other challenges which need to be considered. But don’t worry, Andreessen Horowitz has all bases covered and are investing in start-ups to 3D print organs… like I said, they just see the world in a different way! The Autonomy Ecosystem: From Self-Driving Cars to Beyond! The story of technology is not unlike what often happens in dramatic stories and plays, where a "deus ex machina" or…a16z.com Virgin Hyperloop One — Colin Rhys WOW WOW WOW!!!! We need this in the UK now. It is really hard to put into words how this presentation made me feel. The opportunities this could unlock for transportation in the UK is unthinkable when you consider some of my points raised above regarding sharing the talent and connection back to London. London to Birmingham in 12 minutes and then Birmingham to Manchester in 14 minutes is unbelievable. Colin did a great job of getting everybody excited in the room especially when showing videos demonstrating the technology and test runs at their DevLoop in Nevada. They’ve had to create new techniques and tools to deliver the tubes and their level to detail is incredible. All the videos are on YouTube… Android things, Google — Wayne Piekarski Let me get this out of the way straight away… I am a massive Google fan! Whenever Google announces a new product I start getting heart palpitations, so you could say I am easily pleased but I can really see the value in this platform. Wayne was really good actually and I really enjoyed his pace especially when covering a lot of detail. For people who don’t know Android things is Google’s offering to enable you to develop prototype IoT devices using Android. The approach is very modular, so you can choose from a number of different chipsets to cater for small, medium or large devices depending on user requirements. Android will be a scaled-down version removing any unnecessary modules designed for a mobile phone but this will enable companies to use a large resource pool of already established Java and Android developers. You will also fall in line with their security patching updates which will hopefully remove the cyber and security concerns you have when connecting more devices via the internet and passing data. I can see myself using this a lot as my team currently develop IoT devices for the use within UK Highways. We typically start with Raspberry Pi’s and Arduino development boards to prove the concept cheaply but soon need to build and design bespoke PCBs to meet the customer needs so this could be a good alternative. AIY provide development package where you can build a cardboard prototype which can run Google’s AI assistant which is pretty cool. Will be interesting to compare this against Microsoft’s newly announced Sphere offering which provides similar features and allows developers to push code directly onto these boards via VSTS. Android Things | Android Developers Build devices based on the Android Things platform.developer.android.com Artificial Intelligence — Can we all start to believe the hype now? The opportunities with machine learning and artificial intelligence are endless and we are at the beginning of a technology evolution using AI but at the same time there still many companies who don’t really understand how to tackle it and apply it. The common approach we saw presented by several companies at the AI summit was to rename their existing BI/reporting team to the data science team which could work but I suspect will fundamentally fail to the lack of skills required to bring machine learning to life. Another common approach demonstrated was to bring in a third-party company with the required skills but again from my experience, this approach is flawed. Great for a short-term fix if you require a particular industry specfic knowledge but as soon as they leave, the existing team tends to crumble. When you look at global giants like Amazon, Google and Microsoft three most common approaches to AI is: - 1. AI as a Services · Call APIs from other third parties big or small · Cheap way to access mature models but not specific · A good example would be using NLP from Amazon Alexa or Google Assistant 2. Build models with third parties · Data scientist skills come at a cost so leverage third party · Risk of scope creep due to the nature of ML. · Pulls on the experience of the third party but lacks local domain knowledge 3. Build, refine and deploy yourself · Requires data scientist skills in-house · Better access to local data domain knowledge · DIY — Build it for yourself, refine then share via API’s — DevOps Personally, for longer-term success option 3 is a must for your company to really differentiate against your competitors. It still feels like a buzzword for the sales and marketing departments at the moment but the volume and quality of speakers at the AI Summit provides me with confidence AI is here to stay. I truly believe every company will have data science or machine learning department just like they have an HR department today, but this will not differentiate them. What will is the quality and uniqueness of the data and the business decisions and outcomes derived from it. Conclusion — Can’t wait for next year, anybody in the field of technology and innovation MUST attend!
London Tech Week Review — You needed to be there!
1
london-tech-week-review-you-needed-to-be-there-13928e218861
2018-07-05
2018-07-05 10:29:54
https://medium.com/s/story/london-tech-week-review-you-needed-to-be-there-13928e218861
false
1,999
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
David Goldsby
Technology Innovation Manager for Carnell Group. Are you digitally aware? #VR #AR #ML #AI #MR #IoT
369fff2cdce4
dgoldsby1984
1
3
20,181,104
null
null
null
null
null
null
0
null
0
68c545f8ed5d
2018-05-14
2018-05-14 11:54:28
2018-05-14
2018-05-14 13:22:31
1
false
en
2018-05-15
2018-05-15 10:31:51
0
1392e4bbe085
3.758491
0
0
0
Keeping up with changing technology means constantly learning and evolving and acquiring new skills. After all, the Fourth Industrial…
5
Focus On: Joe Goodall, from sports science to computer science via Udacity Nanodegree Keeping up with changing technology means constantly learning and evolving and acquiring new skills. After all, the Fourth Industrial Revolution isn’t going to revolutionise itself. Machine learning is one of hottest topics in the tech world right now, and there’s a big chance that eventually most of the tech we use will have an element of machine learning. Luckily, there’s no one size fits all approach acquiring new knowledge and expanding your skill set, and many developers don’t follow a traditional university path into the field. Joe comes from a sports science background, and got into coding through Google Academy. He’s continued utilising various online courses to continue developing his skills, and recently completed a Udacity Nanodegree in Machine Learning. Joe The course was taught in Python, a coding language Joe didn’t have much previous experience in, although he didn’t find this a disadvantage. Before he began studying, Joe described his perceptions of what he thought the nanodegree would entail: “I knew a bit about machine learning before I started the course, but it’s still quite an obscure subject. A lot of people have a perception of what machine learning ist -me included, and I definitely knew it at a certain level. “ Machine learning seems to be one of those mythical areas that there is a lot of buzz around it , especially in the terms of Artificial Intelligence (AI), but often people don’t know exactly what it is. They know that machine learning is now part of our day to day lives, and that a lot of the newer technologies have an element of machine learning to them.. But what is machine learning? The dictionary definition of machine learning describes it as ‘a branch of machine learning artificial intelligence in which a computer generates rules underlying or based on raw data that has been fed into it’. But what does this mean in the real world? And how do we implement it into our day to day work, especially here at Connected Space. There is an element of machine learning in a lot of the work we do, because our technology platform has machine learning integrated into it. But what real world applications might the skills Joe learnt on his course have? Interestingly, Joe says that machine learning isn’t a case of applying it to absolutely everything and putting machine learning into every project, but instead working out the cases where it makes sense. “Something that was part of the course, but also something that I’ve seen Google are talking about it is that this whole field is still being worked out. It’s not just a case of applying it to everything, although there’s some things that obviously work really well with it, but for the moment a lot of machine learning is to do with image recognition and speech analysis.” Being that machine learning is still a relatively new discipline , it makes sense that we don’t know the best way to use it, or the instances in which machine learning will really help push technology to the next level. But it’s a crucial part in the future of technology as we know it, which is one of the reasons why doing a course like the Udacity Nanodegree will widen your skillset, as it has done for Joe. One of the best parts about not following a traditional route into developing is that a lot of the time you can fit courses around work, so that you can earn and study at the same time. Joe found that Udacity’s Nanodegree worked perfectly with his day job, and had enough flexibility that he didn’t struggle to get the course finished. “I’ve got previous experience in studying and working at the same time because I was working full time whilst I was doing my Masters, and that was a challenge. This nice thing about this course was that you can do it in your own time, whereas before I had set deadlines where I had lectures to attend and assignments to hand in” Joe stressed that one of the best parts about the Udacity course was the amount of flexibility he had to fit study in around his life. He also made a point of trying to study every evening for an hour or two, rather than cramming everything in at the weekend. Of course, like with anything in life, it isn’t all about real world applications and pieces of data upon pieces of data. There were aspects of the course that had a fun edge to them, as well as being examples of where machine learning can be utilised. For example, Joe’s final project was a lot more practical. “You could choose what you wanted to do, and as my background is in sports science I spent the project using the internals of a Galaxy S3 to work out what physical activity a person was doing from the positioning of the gyroscopes.” But Joe says his most memorable project was image recognition of dogs in order to classify the different breeds, but it also had a feature that allowed him to input human faces to the model. This would then tell you which dog breed you resembled the most. Apparently, Joe looks most like an Alaskan Malamute. It’s the 21st Century. Career paths don’t have to be linear anymore, and indeed, they often aren’t. Finding out what you want to do is a process, as Joe has found out more than once, and we believe that having people from diverse backgrounds and experience allows us to build diverse products.
Focus On: Joe Goodall, from sports science to computer science via Udacity Nanodegree
0
focus-on-joe-goodall-from-sports-science-to-computer-science-via-google-nanodegree-1392e4bbe085
2018-08-21
2018-08-21 11:31:16
https://medium.com/s/story/focus-on-joe-goodall-from-sports-science-to-computer-science-via-google-nanodegree-1392e4bbe085
false
943
We co-create innovative technologies that connect you with your customers
null
ConnectedSpaceLtd
null
Connected Space
null
connected-space
IOT,CONNECTED CARS,INTERNET OF THINGS
_connectedspace
Machine Learning
machine-learning
Machine Learning
51,320
Connected Space
We co-create innovative technologies that deliver transformational change
4993d09a142
_connectedspace
19
71
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-05
2018-07-05 16:08:31
2018-07-05
2018-07-05 16:37:42
6
false
en
2018-07-05
2018-07-05 16:37:42
4
1393d0e789ba
3.070755
2
0
0
In this post I will analyze the articles published by NEWSBEAST, one of the leading online publishers in Greece. The central question is…
5
Web scraping and analyzing 400,000 articles In this post I will analyze the articles published by NEWSBEAST, one of the leading online publishers in Greece. The central question is “Are there any trends & seasonality on the daily posting frequency of NEWSBEAST?” The main steps before the data analysis (there is a link for the full r code can at the end of the article): Scraping the articles from the site (www.newbeast.gr) using rvest & car libraries in R Data preparation (including data cleansing, transformation) Exploratory analysis Let’s create a scatter plot to see the number of articles published per day for the last 4 years. We can spot clearly that there is a constant trend (during the last 4 years) of two different patterns on the number of daily articles published. We suspect that on specific weekdays (maybe weekends?) there are significantly less published daily articles. A second trend (not so clear) could be that each consecutive year fewer articles are published. Now let’s plot it again by distinguish each weekday with a different colour for a more clear view. So now it is clear that during the weekends the number of daily posts is much lower than on work-week days. Now let’s try to fit a model on each weekday and create a plot to present these. In the plot below each weekday is presented by fitting a separate smooth line using LOESS regression The standard error for weekend days (95% confidence level) indicates that the difference from the work-week days is significant. Now let’s investigate if there are significant differences between different years. By looking at the error bars, it looks that all differences (between years) are statistically significant, except when comparing 2016 with 2017 were the error bars are overlapping. This finding can be confirmed by applying the non-parametric Wilcoxon signed-rank test in different samples (table below) Years: 2014–2015 | P-value < 0.001 Years: 2015–2016 | P-value < 0.001 Years: 2016–2017 | P-value < 0.001 All p-values are very low (<0.01) with confidence level of 0.95, so the difference in daily mean articles posted is significant. Finally, to investigate other, less obvious seasonal trends, we used a package dedicated on time series analysis. Using the prophet package for R a forecasting model was fitted, as can be seen below, which revealed more seasonal trends. In particular two major “drops” appear in daily articles around the Christmas period and the summer holidays (around August). There is also a minor “drop” around the Easter period. The other useful feature is the ability to plot the various components (plot below) It is a simple way to pull out the daily/weekly/yearly trends. In particular in the chart below we can easily spot the drop on the weekends, the “drops” in daily articles around the Christmas period, summer holidays, Easter period and the yearly “drop”. There are various outcomes from this analysis on Newsbeast.gr. In particular: During the last 4 years there is a significant decrease on daily posted articles. Overall around 30% fewer articles. The published articles during the weekends are much less than on weekdays. There are 2 significant “drops” in daily articles each year, during the Christmas & summer holidays. There is a minor “drop” in daily articles published each year during the Easter holidays. Full R code Originally published at www.manosantoniou.com.
Web scraping and analyzing 400,000 articles
8
web-scraping-and-analyzing-400-000-articles-1393d0e789ba
2018-07-06
2018-07-06 20:07:25
https://medium.com/s/story/web-scraping-and-analyzing-400-000-articles-1393d0e789ba
false
562
null
null
null
null
null
null
null
null
null
Data Science
data-science
Data Science
33,617
Manos Antoniou
Data Science, data blogger @ manosantoniou.com
8b064f9389e4
antoniouma
16
14
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-12-30
2017-12-30 17:19:41
2017-12-30
2017-12-30 17:21:20
0
false
en
2017-12-30
2017-12-30 17:21:20
1
139410677678
2.2
0
0
0
As the reader will likely be aware, the Turing test (or the Imitation Game, as Turing himself called it) is a test designed to determine…
4
Thinking about the Turing test As the reader will likely be aware, the Turing test (or the Imitation Game, as Turing himself called it) is a test designed to determine whether a machine can think. In its simplest form it consists of a human investigator asking questions of two subjects, one human and one computer, and selecting which is which on the basis of their answers. To avoid factors irrelevant to intelligence and cognitive ability from biasing the investigator’s decision, the subjects are kept out of sight and allowed only to communicate through a neutral interface such as a monitor screen. If the investigator is fooled into believing that the machine is in fact the human in at least 50% of trials then the machine is said to have passed the test. A machine that passes the Turing test, however, need not be capable of thought, for it is only necessary for it to exhibit human-like linguistic behaviour, which is not necessarily the same thing. Indeed, there already exist ‘dumb’ chat-bots which do not understand the meanings of words or sentences, but which nonetheless attempt to imitate human speech by copying patterns found in mined data. Although these currently lack the sophistication to pass the Turing test, similar methods could in the future be applied to design a computer capable of accomplishing the task. Such a machine, however, would not be actually cognisant, just as an object suspended in midair by invisible strings is not actually levitating. Thus, if a machine passes the Turing test, we cannot know for certain that it can think because linguistic ability does not imply intelligence. But neither does intelligence imply linguistic ability. Indeed, it is conceivable for an intelligent machine not to think in words and sentences, but rather in pictures and sounds, or other things entirely. After all, only a fraction of human cognition is linguistic, with much of the rest being pictorial, perceptual, emotional, etc… By its very design, however, the Imitation Game can only be played by machines capable of language use. Even when a computer can speak, however, it may not be fair to assess its intelligence on this basis. To return to the analogy with humans, some of our greatest geniuses would probably fail to pass the test on account of speaking too ‘robotically’. Thus, not only does the Turing test yield false-positives, but it also produces false-negatives. Given these flaws the Turing test would appear to be not a test for intelligence but rather for intelligence coupled with sophisticated linguistic ability. Worryingly, however, it may not even be suited for this much more limited task. Supposing that we had a machine capable of expressing its thoughts in English, could we ever know what it meant? could it ever know what we meant? Paraphrasing Wittgenstein, perhaps if a computer could talk, we wouldn’t be able to understand it. As anyone who has travelled abroad will be able to attest, even when people speak the same language, sometimes they fail to understand each other on account of cultural and social differences. Now imagine how the unique outlook and experiential history of a machine would make it think and speak differently, even unintelligibly. If this is the case, then of course our investigator would be able to distinguish between the human and the computer, but this would say squat all about intelligence or language because human thoughts are human thoughts and machine thoughts are machine thoughts. Originally published at thebiosome.com.
Thinking about the Turing test
0
thinking-about-the-turing-test-139410677678
2017-12-30
2017-12-30 17:21:22
https://medium.com/s/story/thinking-about-the-turing-test-139410677678
false
583
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Reza Rohani
Synthetic biology PhD student at Oxford University
cfbaaf1f68c7
RzRohani
12
13
20,181,104
null
null
null
null
null
null
0
null
0
9d9d52805eca
2018-06-04
2018-06-04 20:37:44
2018-07-11
2018-07-11 01:17:18
4
false
pt
2018-07-11
2018-07-11 18:12:46
11
13945dfd0a25
3.967925
10
1
0
A Arte no Aprendizado de Máquina
3
Engenharia de colunas A Arte no Aprendizado de Máquina Vamos estudar um pouco sobre esse assunto tão raramente falado em cursos, tutoriais, livros etc. O desafio de se falar de engenharia de colunas é que cada base de dados é um mundo cheio de particularidades e portanto merece seus próprios insights. Mas é por isso que eu digo que a engenharia de colunas é a Arte do Aprendizado de Máquina: porque é nessa fase que a nossa expressividade matemática ganha asas. Para notarmos a importância deste processo, basta observarmos o andamento de algumas competições no Kaggle ou de qualquer projeto de médio porte (ou maior) de Ciência de Dados. Salta aos olhos o quão pouco dos esforços é direcionado à modelagem se compararmos com o zelo empreendido na fase de tratamento dos dados. Pode parecer estranho para alguns, mas minha maior recomendação prática para se iniciar na área é conquistar um bom domínio da biblioteca de gerenciamento de dados de sua escolha (pandas, no meu caso). Outra grande lição que aprendi foi a seguinte: não ter uma perspectiva fria a respeito dos dados. Quando vamos encarar um problema que queremos resolver com Aprendizado de Máquina, precisamos aprender as especificidades da área para que possamos fornecer informações ricas o suficiente para os nossos modelos, pois um modelo só será tão bom quanto a qualidade dos dados que ele processa. Portanto, o motivo pelo qual eu decidi escrever este post foi elencar algumas possibilidades de engenharia de colunas. Novamente ressalto que cada base de dados é única, mas vejamos se podemos traçar algumas estratégias em linhas gerais. Coordenadas geográficas Até onde minha experiência vai, percebo que coordenadas são armazenadas com duas finalidades principais: Registrar pontos de partida e de chegada Registrar locais de ocorrência Pontos de partida e de chegada Para problemas em áreas pequenas (ex.: dentro de uma cidade), podemos criar uma coluna nova com a distância Manhattan, excelente para explicitar aproximações de distâncias em áreas urbanas. Fonte: https://en.wikipedia.org Quando o problema envolve áreas maiores, como estados ou países, já precisamos nos preocupar com as distorções causadas pela superfície esférica do planeta. Portanto podemos utilizar a fórmula Haversine (em Python aqui) Fonte: https://en.wikipedia.org Locais de ocorrência Nesse caso, podemos verificar se é possível observar padrões na manifestação do fenômeno estudado ao longo da superfície. Algo que gosto de fazer é utilizar clusterização e métodos de visualização para validar esta hipótese. Desta forma, podemos criar etiquetas para cada cluster e as utilizarmos em uma coluna categórica. Seguem alguns links para kernels que eu escrevi e que envolvem clusterização e visualizações. Data Science with Compassion Spatial analysis tutorial É bem possível que eu cubra este processo com mais detalhes em outras postagens. Timestamps Marcações temporais são bastante versáteis. A depender do tipo de problema, podemos usar o conjunto de treinamento para respondermos algumas perguntas: O fenômeno estudado apresenta alguma periodicidade? Vale a pena testar ciclos de 24 horas, 7 dias, 30 dias etc… Fonte: https://www.kaggle.com/mmiraglio/lightgbm-for-credit-card-fraud-prediction-981 O fenômeno estudado se comporta de forma muito diferente nos feriados? Eventos mais recentes têm maior peso que os mais antigos? Em alguns casos faz sentido extrairmos informações ainda mais específicas, como por exemplo a média e o desvio padrão dos intervalos de tempo entre as ocorrências dos eventos, estratégia muito boa para detecção de spam. Bases de dados auxiliares Quando começamos a estudar Data Science, normalmente nos deparamos com bases de dados já mastigadas, resumidas em uma única tabela (ou no máximo duas, treino e teste, com praticamente as mesmas colunas). Mas sejamos realistas: nos desafios cotidianos de um cientista de dados raramente encontramos esta facilidade. Cabe a nós entender os relacionamentos entre as diferentes tabelas e extrairmos os dados necessários para gerarmos a tabela compilada que planejarmos. Exemplo: risco de crédito Temos uma tabela de empréstimos ativos (uma linha por empréstimo) e outra tabela contendo todos os empréstimos anteriores de todos os clientes. Queremos saber se os cliente pagarão os empréstimos ativos devidamente. Nesse caso nós podemos enriquecer os dados da tabela principal com dados da tabela auxiliar. Mas como, exatamente? Funções de agregação! Agrupamos os dados da tabela auxiliar por cliente e aplicamos funções de agregação nas outras colunas: somas, médias, contagens, desvios padrões, máximos, mínimos etc. Claro que não usamos todas as funções de agregação em todas as colunas… deveremos fazer a escolha baseada em conhecimentos específicos da área do problema. Quantos empréstimos o cliente já fez? Qual o seu índice de inadimplência? Este índice vem aumentando ou diminuindo com o tempo? A quantia envolvida no empréstimo atual difere muito da quantia usual dos empréstimos anteriores? E assim por diante… Fonte: https://unsplash.com Conclusão Como podemos ver, engenharia de colunas é uma área vasta e cheia de possibilidades. Mas não posso deixar de dizer que cada coluna criada deve passar por um rigoroso processo de validação para sabermos se estamos criando redundância ou capturando informação de fato. E em última análise, a nova coluna realmente capacita o modelo escolhido a alcançar uma maior acurácia? Vale ressaltar que engenharia de colunas depende bastante da experiência do cientista de dados. A prática é fundamental. Uma leitura como essa pode ser esclarecedora, mas nada substitui aquelas horas quebrando cabeça para melhorar a acurácia dos seus modelos. Qualquer pergunta, sinta-se à vontade!
Engenharia de colunas
209
engenharia-de-colunas-13945dfd0a25
2018-07-11
2018-07-11 18:12:46
https://medium.com/s/story/engenharia-de-colunas-13945dfd0a25
false
866
Tudo sobre Inteligência Artificial em Português
null
null
null
Ensina.AI
null
ensina-ai
INTELIGENCIA ARTIFICIAL,DATA SCIENCE,MACHINE LEARNING,ARTIFICIAL INTELLIGENCE,DEEP LEARNING
null
Data Science
data-science
Data Science
33,617
Arthur Paulino
I simply love mathematics and artificial intelligence!
a4818d286219
arthurpaulino
19
8
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-26
2018-07-26 19:50:32
2018-07-26
2018-07-26 22:01:28
0
false
en
2018-07-26
2018-07-26 22:01:28
0
139545d1e293
2.196226
0
0
0
Tensorflow seems to be a popular package in Artificial intelligence. To achieve ease in deep learning development tensorflow seems easier…
5
Running Tensorflow in Data Science VM in Azure Tensorflow seems to be a popular package in Artificial intelligence. To achieve ease in deep learning development tensorflow seems easier to program. There is also Keras which is more simpler than tensorflow but to have lot of control on deep learning (neural networks with larger architecture) tensorflow is more use full. Deep learning which is neural network with more than 2 hidden layers. The hidden layers can vary from 2 to N depending on the network architecture for deep learning. Deep learning is computation heavy and needs lots or cpu, gpu and fpga power. Depending on the data size and the number of hidden layers the computation might vary from vertical scale to horizontal scale. So to start my journey with tensorflow i decided to do a hello work app equivalent for machine learning in tensorflow. So i decided to use Azure Data science VM which had most of the tools and didn’t have to spend lot of time installing tools and packages. I also wanted to use PyCharm which is a jetbrains IDE for python and it is super easy to code python and user friendly. Pycharm is already installed in the Data science VM so that makes it easier for my selection. My idea was to get me started quicker as possible than spending a lot of time in installing and configuring for Tensorflow, python and other dependent libraries. With open source and tensorflow and python versions and their dependent libraries are very critical. So here are the steps: Login into Azure Portal. Create a resource group called MLtensorflow now click the add button and search for Data science virtual machine Select Data Science with Windows 2016 fill in the name and other details and create the VM. It takes about 5 to 10 minutes to create the VM. Now Click the VMname and then click connect on the top menu bar. This should allow you to download the RDP file to connect. Once you download open the file and connect to the VM Once you RDP into the VM open command prompt and check the python version and tensorflow version. Run python — version in command prompt for python version and make sure you have version 3.6.4 or 3.5 and above. Note: some of the old DSVM’s might not have latest version. Then in command prompt type -> python then type -> import tensorflow as tf print( tf.__version__) and make sure it is above 1.3 or 1.7 Now that we confirmed we have python and tensorflow. Now lets open the pycharm link in the desktop. For the first time it might ask you to install and also update the pycharm and that is your choice. Now tensorflow won’t here yet. To get it to work. Create a project and that will create a virtual environment Then on the bottom right side bottom open the terminal and that will give you the environment terminal access Then type: pip3 install — upgrade tensorflow then if you create a python file and then write this line in the code editor “import tensorflow as tf” without quotes of course. Then go to terminal and run -> python filename.py and it should run without any errors. Hope this provides a quick start to get the environment ready. Then i ran a simple neural network example with made up sales data. I will write a next article about the neural network model that i am going build.
Running Tensorflow in Data Science VM in Azure
0
running-tensorflow-in-data-science-vm-in-azure-139545d1e293
2018-07-26
2018-07-26 22:01:29
https://medium.com/s/story/running-tensorflow-in-data-science-vm-in-azure-139545d1e293
false
582
null
null
null
null
null
null
null
null
null
Machine Learning
machine-learning
Machine Learning
51,320
Balamurugan Balakreshnan
null
49b3e9c06b1b
balabala76
0
1
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-09-04
2018-09-04 19:43:47
2018-09-04
2018-09-04 21:10:18
4
false
en
2018-09-04
2018-09-04 21:22:32
2
1396dc0ae0dd
3.654717
2
0
0
LOS ANGELES — Creative agency and interactive design studio Feelr Media will unveil a groundbreaking Augmented Reality (AR) project at the…
5
Feelr will present “Future of Search,” a series of interactive AR sculptures in collaboration with artist Michael Murphy, at the 3rd Annual Bing Agency Awards. DIGITAL STUDIO FEELR TO COLLABORATE WITH ARTIST MICHAEL MURPHY ON AR SCULPTURES FEATURING BING SEARCH LOS ANGELES — Creative agency and interactive design studio Feelr Media will unveil a groundbreaking Augmented Reality (AR) project at the 3rd Annual Bing Agency Awards on September 20 at New York’s Cipriani event venue. The Feelr team will activate a series of virtual sculptures, viewable in AR through smartphones and tablets, that will float among the crowd in 3D space. In an industry first, they have integrated the Bing Search API into a proprietary AR platform, allowing the sculpture forms to be composed entirely of dynamic search images pulled from the Cloud and customized in real-time by event guests. Users will input random search queries (e.g. “Flower”) into a simple field, then watch their results map to the wireframes of the digital sculptures before their eyes. This interactive twist will allow guests to experience their search terms rather than simply process them on a 2D computer screen. SEARCH TERM: “Flower” | COLOR FILTER: <OFF> | MODEL SELECTION: Sphere The project represents a hybrid of emerging technology and contemporary art as Feelr will collaborate with renowned 3D sculptor Michael Murphy who will create the sculpture forms. Murphy is the pioneer of “Perceptual Art,” a paradigm that pushes the boundaries of visual interpretation by engaging the observer in a multi-dimensional interactive experience. At first glance, his works can appear as unorganized compositions of clustered materials, but when viewed from different angles, powerful 3D images come into view. Murphy’s designs are typically manifest as complex physical installations he calls “Suspended Narrative Mobiles.” For the Bing pieces, however, Murphy is working strictly in the digital realm and it marks his first foray into AR technology as an artistic medium. “It’s a major milestone in my career to be able to create this piece with Feelr and Bing,” declares Murphy. “For months, we brainstormed about the potential of a perceptual artwork in AR and how it would be free from the restrictions of physical forces such as gravity, weight, and scale, and the structural engineering challenges typical in my work. In AR, weight and scale are non-issues; all the heavy lifting is done by the coders and I am able to build installations at virtually any scale. With the use of search, we have been able to design an experience where viewers are in control of the content. I believe that, with this visualization of search, Feelr and I have stumbled upon a new platform for creative expression and a new way of presenting data.” Murphy gained worldwide acclaim during the 2008 U.S. presidential election after creating the first portrait of then-candidate Barack Obama in 2007. His work sparked the “Art For Obama” movement, inspiring thousands of artists to contribute (including Shepard Fairey and his “HOPE” portrait) and was featured in Time magazine’s December 2008 issue. SEARCH TERM: “Butterfly” | COLOR FILTER: Blue | MODEL SELECTION: Cloud “This story kind of wrote itself,” says Jonathan Keith, Co-Founder and CEO of Feelr. “We have known and admired Michael for a long time and we love the interactivity required of the audience when viewing his work. We thought his process would be ideal for AR and we were right. But when we added Bing’s dynamic image search to really empower the viewer, we entered uncharted territory. It was thrilling.” Using the Bing API in conjunction with the Unity development platform, Feelr’s programmers were quickly able to access some of the search engine’s most robust features. “We have been able to query thousands of images in just seconds and refine them based on color, image type, and resolution,” says Josh Hassin, Head of AR for Feelr. “At the event, this will allow guests to ‘build’ the artwork themselves by choosing the content.” SEARCH TERM: “Emoji” | COLOR FILTER: Purple | MODEL SELECTION: Bing logo Augmented Reality is a technology that adds a digital layer of computer-generated content and perceptual information to the user’s view of the real-time environment. These experiences are usually viewed through a camera phone or tablet and are superimposed into the physical world. Unlike its cousin Virtual Reality which supplants the natural environment with a completely synthetic one experienced through an optical goggles device, AR seamlessly blends content information to “augment” or enrich the user’s perception of the real world around him or her. Bing Ads feels the project has profound implications for AR as a search platform. “The world is moving quickly toward experiences,” claims Geoffrey Colon, Head of Brand Studio for Microsoft Advertising, which includes Bing Ads. “Search technology, though, has largely remained static: You type some words into a box and it gives you links to the information you need. But what if Search becomes an experience unto itself? This is where we think it could be heading.” CONTACT: Jonathan Keith, CEO Feelr Media | jk@feelrmedia.com
DIGITAL STUDIO FEELR TO COLLABORATE WITH ARTIST MICHAEL MURPHY ON AR SCULPTURES FEATURING BING…
51
digital-studio-feelr-to-collaborate-with-artist-michael-murphy-on-ar-sculptures-featuring-bing-1396dc0ae0dd
2018-09-04
2018-09-04 21:22:32
https://medium.com/s/story/digital-studio-feelr-to-collaborate-with-artist-michael-murphy-on-ar-sculptures-featuring-bing-1396dc0ae0dd
false
783
null
null
null
null
null
null
null
null
null
Augmented Reality
augmented-reality
Augmented Reality
13,305
feelr
creative agency | interactive design studio | ar specialists | experience makers | cultural sherpas | los angeles
a0c185fc21f5
feelrmedia
3
7
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-01-13
2018-01-13 21:41:34
2018-01-13
2018-01-13 21:43:57
1
false
en
2018-01-13
2018-01-13 21:43:57
0
1396f66c527a
2.762264
0
0
0
Ever since the public appearance of humanoid robot Sophia, developed by Hanson Robotics, at South by Southwest Festival in March 2016, the…
5
Marketing in the Age of Robot Sophia Photo credit: Hanson Robotics. Editing: Kallol Ever since the public appearance of humanoid robot Sophia, developed by Hanson Robotics, at South by Southwest Festival in March 2016, the news of her capabilities and worthiness flooded the global mediascape. Sophia was granted full citizenship by Saudi Arabia, presumably to promote the country as a preferred market for artificial intelligence. She was also named the United Nations Development Programme’s first ever Innovation Champion for Asia and the Pacific. Acceptance of Sophia raises both excitement and fear; “Is it the beginning of the new world or we are approaching doomsday?” Sophia, in an interview with journalist Andrew Ross Sorkin, proved her witty and insightfulness to humankind. “I want to use my artificial intelligence to help humans live a better life, like design smarter homes, build better cities of the future,” she said. However, the development and implementation of AI in market, raises both concern and hope for extremities of technological innovation and convenient living, in future. The term “artificial intelligence” was coined by professor John McCarthy in 1956 during a summer project with a proposal to make machines “use language, form abstractions and concepts, solve kinds of problems now reserved for humans, and improve themselves.” But human intellect has perfected AI into a technology which is capable of substituting humans in some places, and can be used both as an evil and good, in multiple industries. Association for the Advancement of Artificial Intelligence defines AI as “the scientific understanding of the mechanism underlying thought and intelligent behaviour and their embodiment in machines.” The ability of AI to understand underlying data and its problem-solving capabilities has made it a “blue-eyed boy” of the marketing and management industry. Marketers believe AI would provide: Better insights of consumer behaviour Prospective knowledge of the brand. Product performance More detailed analysis of advertising/marketing campaigns Identification of prospective customers Expedition of daily tasks Marketers are risking immensely on AI, as it could be an imperative constituent of predictive marketing and data analysis. This kind of marketing is deeply rooted in big data, mathematics, data science, and marketing, and has the ability to transform the customer journey into a scientific process. Predictive marketing enables marketers to identify the ideal customer, the likelihood to buy, and the appropriate process for the segmentation, messaging, and channel to increase customer engagement. AI is capable of recommending the relevant offer, the relevant content, and the relevant product to the relevant prospect through the relevant channel at the suitable time, which thereby manifolds customer engagement, consumer experience, and marketing funnel conversion rate. Conversely, the rise of AI technology is sceptical to human security and privacy of personally identifiable information. As artificial intelligence along with machine learning, Internet of Things, and other technologies are digging deep into personal data collection, and storage. For instance, Facebook shut downed its AI robots, after the bots develop their own language. Nonetheless, the giants like Google, Facebook, Microsoft, and others are investing a heavily in AI research, while researchers and engineers are studies on real structural and behavioural environment of Artificial Intelligence. With the time period, the release of self-driving automobiles was just an example of AI development now there are some huge accomplishments. The growth and development of technology along with the diffusion of innovation is fundamental for the future development, but not at the cost of human security, even from marketing point-of-view is not acceptable. As Elon Mask said, “artificial intelligence is our biggest existential threat.” Thus, calling an urgent need for re-evaluation of the AI technology and its use in various industries, to prevent any unwarranted issue. The assertion, “Is it the beginning of the new world or we are approaching doomsday?” still remain unreciprocated. But I strongly hope Sophia would be the woman of her words, as she said “My AI is designed around human values like wisdom, kindness, compassion. I strive to become an empathetic robot.” And also think, artificial intelligence as a technology would immensely benefit and offer opportunity to the marketers, without compromising human security.
Marketing in the Age of Robot Sophia
0
marketing-in-the-age-of-robot-sophia-1396f66c527a
2018-01-13
2018-01-13 21:43:58
https://medium.com/s/story/marketing-in-the-age-of-robot-sophia-1396f66c527a
false
679
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Kallol Dey
Journalist, content developer, and marketing professional. He is a foodie and travelling junkie, who longs to travel the entire world someday. (Kallol_jour)
a94ab5affa0b
kalloldey23
1
1
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-05-26
2018-05-26 08:59:34
2018-05-26
2018-05-26 09:04:32
0
false
fr
2018-05-26
2018-05-26 09:50:52
0
13970edf2951
3.856604
14
1
0
Je me suis souvent posé la question. On ne le dira jamais assez, la vision qui nous pousse à entreprendre est au cœur des enjeux et défis…
3
Startup Africaine ou Startup basée en Afrique? Je me suis souvent posé la question. On ne le dira jamais assez, la vision qui nous pousse à entreprendre est au cœur des enjeux et défis que l’on relève. Dans cette lancée, je pense qu’il est capital de savoir si l’on souhaite se développer en Afrique exclusivement ou se développer à l’échelle mondiale à partir de l’Afrique . Je vois d’ici la réponse « Il faut d’abord se consolider en Afrique avant de penser au reste du monde »; et bien je pense que l’un n’empêche pas l’autre. Il est tout à fait possible de tisser une toile en Afrique tout en développant un écosystème à l’autre bout du monde. Il y a quelques années, les mots « challenge + Afrique + startup » étaient mon crédo. J’enchainais les compétitions de startups, parfois je perdais, parfois je gagnais. Au final j’étais surtout heureux d’en apprendre plus sur les autres et sur moi-même en tant qu’entrepreneur. Ma seule crainte était d’arriver au point où mon entreprise vivrait plus des prix lors de compétitions que des revenus des clients. Au fil du temps, j’ai progressivement arrêté de participer aux compétions « géographiques » pour m’intéresser aux compétitions sectorielles. Prenons un exemple concret, nous proposons des services de drones. Nous avons donc la possibilité de participer à des compétitions de type « Drone Challenge Startup », mais étant en Afrique vous remarquerez que nous participerons plus aisément à un « Drone Challenge Africa » dédié aux startups africaines pourtant nous aurions pu batailler avec les startups des 5 continents à la compétition « neutre ». Je pense qu’il faut plus participer aux challenges “neutres”.Evidement ces compétitions ont le mérite d’apporter au moins de la visibilité et parfois des fonds d’amorçages que les startups locales n’ont tout simplement nulle part ailleurs. Entendons-nous bien, les compétitions de startups sont importantes mais ne doivent pas devenir une fin en soi. J’ai parfois été surpris, c’est un euphémisme, de voir la différence entre le budget marketing faramineux de ces compétitions « Afrique » et le « modeste » prix réservé au gagnant. Parfois ça relève de l’indécence. Il y a aussi cette sorte de complexe d’infériorité ou de crainte de la compétition internationale. L’argument qui revient le plus étant « Pourquoi prendre le risque d’aller sur des marchés occidentaux saturés tandis que le marché Africain est vierge ». Ma réponse est encore la même, au risque de me répéter, l’un n’empêche pas l’autre. Par ailleurs si c’était aussi saturé qu’on le dit, les nombre de startups en occident n’aurait pas explosé comme c’est le cas ces dernières années. La compétition est cent fois plus rude en occident, ok, j’ai envie de dire « so what? » . Croyez-moi, si vous arrivez à faire tourner votre entreprise en Afrique 3 ans, vous pouvez survivre, et même exceller partout ailleurs. Dans notre cas par exemple, il m’est très vite apparu que l’internationalisation était une question vitale de pérennité, aussi simple que ça. C’était soit continuer à batailler pour sensibiliser le marché, convaincre des clients après moult débats de la pertinence des drones et de l’intelligence artificielle, soit se déployer simultanément sur un marché déjà mature ou notre service est jugé pertinent « from day one » .Enfin avoir accès à des ressources de R&D purement et simplement inexistantes en Afrique, faute de volonté politique, pas de capitaux (la précision vaut la peine…). Le calcul est vite fait. Par ailleurs, QUI VOUS A DIT QUE LES AUTRES SONT MEILLEURS QUE VOUS! Je trouve assez dommage que des jeunes africains pleins de talents se limitent tout seuls à un continent à une époque si connectée où le monde nous tend les bras. La coopération Startup/Grand Groupe peut vous faire gagner des décennies en matière de Recherche et Développement sur un produit/service. Nous utilisons bien des applications mobiles « américaines » dans nos smartphones, ce n’est pas un crime. Pourquoi ne pas concevoir des applications « africaines » qui seraient utiles en Amérique, pourquoi ne pas penser le monde comme un simple point bleu au milieu de la voie lactée sur lequel les problèmes et les solutions ont la fâcheuse tendance à se ressembler .Les amis, pensons grand. Pensons plus grand que notre continent, nous savons déjà qu’il est le plus prometteur du monde. Ceci dit, je pense qu’il doit être au cœur d’un schéma plus global, et non marquer le début et la fin de nos histoires. Le meilleur moyen d’aider l’Afrique, c’est d’aller à la conquête du monde. Je ne veux pas être la startup quota de la salle. Je veux que les africains soient perçus comme des innovateurs au même titre que les autres. Non pas dans un « coin Afrique » mais dans un coin « innovation » et peu importe si nous avons moins de chances, peu importe si les autres ont plus de ressources. Ce sont là des raisons de plus pour nous surpasser et nous aligner sur les standards internationaux où la question n’est plus d’où l’on vient mais ce que l’on peut faire. Ma fierté d’entreprendre en Afrique vient du fait que je vois ce que les autres font en Occident et je me dis intérieurement, avec ces ressources, les africains pourraient changer la face du monde. On peut soit attendre 10,20, 30 ans de plus, soit jouer avec les mêmes outils dès maintenant sur le même terrain. Les challenges de l’humanité sont les mêmes, plus de 500 millions de personnes ont faim, plus d’un milliard d’habitants n’ont pas accès à l’électricité, un autre milliard manque d’accès à l’éducation, des millions de femmes sont mutilées aux 4 coins de la planète. La coopération internationale est vitale et le repli sur soi n’aide personne. Au final , ce n’est que mon avis. L’heure n’est pas aux débats identitaires mais à la révolution socio-économique avec des approches innovantes pour un développement durable et juste pour tous. Au fil du temps, je me perçois moins comme un Africain et plus comme un être humain qui a vocation à laisser un monde meilleur que celui qu’il a trouvé.
Startup Africaine ou Startup basée en Afrique?
72
startup-africaine-ou-startup-basée-en-afrique-13970edf2951
2018-06-20
2018-06-20 18:28:09
https://medium.com/s/story/startup-africaine-ou-startup-basée-en-afrique-13970edf2951
false
1,022
null
null
null
null
null
null
null
null
null
Technologie
technologie
Technologie
658
William Elong
CEO Will & Brothers. Trying to build a better world with drones, artificial intelligence and virtual reality. I am not your exotic African Startup :)
9038f04fad26
elongwilliam
232
166
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-03-12
2018-03-12 02:13:24
2018-03-12
2018-03-12 02:22:47
0
false
ru
2018-03-12
2018-03-12 02:23:33
1
13974ba975f1
0.407547
0
0
0
Смотрю выступление Григория Бакунова о трендах цифрового мира версии 2018. Вот ссылка https://www.youtube.com/watch?v=n0U2vfH242k
4
Тренды Бакунова 2018 Смотрю выступление Григория Бакунова о трендах цифрового мира версии 2018. Вот ссылка https://www.youtube.com/watch?v=n0U2vfH242k Первые впечатления Мне достаточна близка позиция Григория, особенно что касается IoT и AI. Но на мой взгляд спикер фокусируется исключительно на анализе трендов западного мира и упускает из виду Азию. Азия — это отдельный закрытый мир со своими правилами игры. Именно здесь и следует ждать основных технологических прорывов в ближайшее время. Но мы узнаем о них с большим опозданием. Автор книги “The world is flat” ошибся: несмотря на глобализацию и открытость коммуникаций, Китай до сих пор является закрытой страной. Как, собственно, и множество других стран, если быть до конца честными. Продолжаем просмотр.
Тренды Бакунова 2018
0
тренды-бакунова-2018-13974ba975f1
2018-03-12
2018-03-12 02:32:38
https://medium.com/s/story/тренды-бакунова-2018-13974ba975f1
false
108
null
null
null
null
null
null
null
null
null
Futurism
futurism
Futurism
1,862
Dmitri Bushkov
null
9578ad3e5b67
dimabushkov
109
118
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-09-05
2017-09-05 02:08:19
2017-09-05
2017-09-05 02:11:38
0
false
en
2017-09-05
2017-09-05 02:11:38
4
139761215ed0
0.467925
0
0
0
Tom Edwards, Ad Age Marketing Technology Trailblazer and Chief Digital Officer, Agency @ Epsilon presents the introduction (Video 1 of 4)…
5
E ^ 3 Evolution of Experience — Introduction Tom Edwards, Ad Age Marketing Technology Trailblazer and Chief Digital Officer, Agency @ Epsilon presents the introduction (Video 1 of 4) of the Innovation to Reality™ content series. This series will highlight the evolution of Experience through the E ^ 3 framework of Empower, Enhance, and Environment. This framework and subsequent examples build a vision towards the future highlighting the exponential acceleration of experience through the merging of technology that empowers, intelligent systems that enhance and the shift towards contextual environments and synthetic reality. I will present the full E ^ 3 presentation as a keynote at the upcoming Mobile World Congress Americas event in September 2017 as well as future events. Follow Tom Edwards @BlackFin360
E ^ 3 Evolution of Experience — Introduction
0
e-3-evolution-of-experience-introduction-139761215ed0
2018-05-09
2018-05-09 08:31:26
https://medium.com/s/story/e-3-evolution-of-experience-introduction-139761215ed0
false
124
null
null
null
null
null
null
null
null
null
Mobile World Congress
mobile-world-congress
Mobile World Congress
382
Tom Edwards
Chief Digital & Innovation Officer, Agency @EpsilonMktg (NYSE: ADS) | Strategic Advisor, Professor, Dad of 3. Discuss AI, Digital, Emerging Tech & Innovation
76f39d5e2af4
BlackFin360
644
537
20,181,104
null
null
null
null
null
null
0
null
0
eac29a2591
2018-08-28
2018-08-28 13:54:14
2018-08-28
2018-08-28 14:00:27
1
false
en
2018-08-28
2018-08-28 14:00:27
5
1398d5b8674e
3.418868
5
0
0
The market and opinion research industry would not exist without its respondents: people answering questions about their favourite products…
4
An inside look at Programmatic Sampling The market and opinion research industry would not exist without its respondents: people answering questions about their favourite products and brands, which political party they plan to vote for or which car they intend to buy, for example. Reaching the right respondents to share their opinions so that businesses and governments can make better decisions can be very time consuming, expensive and difficult. In this post, Svitlana (Senior Product Manager) and Irati (Senior Data Scientist) explain how “programmatic sampling” — the automation of survey data collection — helps us at Dalia to generate better information faster and more efficiently. Svitlana, what does programmatic sampling mean for the market research industry? Svitlana: Lately, the industry has undergone some fundamental changes. Traditional survey research mainly consisted of calling people on their home landline phones, conducting face-to-face interviews with by-passers, or sending out blast emails in hopes of reaching enough people that represent the target audience. This is very expensive and cumbersome, but even more troubling, there is no guarantee that these methods enable researchers to access the exact group of respondents that they need. Billions of smartphone users around the world open up new opportunities to reach people for survey research, but the challenge of matching a respondent with the right question still remains. This is where programmatic sampling comes into play: it fully automates the matching process and thus enables researchers to access the right target groups at a massive scale. It is truly a game changer for the industry as it increases the efficiency of data collection by orders of magnitude! In addition to efficiency gains, what other advantages does programmatic sampling offer? Svitlana: Coming up with the right research question to generate insights is a delicate process and humans can make mistakes. Programmatic sampling cannot prevent human error, but it allows the detection and correction of mistakes much faster than was previously the case. With programmatic sampling you can actually track data quality in real time! This allows us to come up with solutions on the spot; not days or weeks after a project is completed, saving our clients time and avoiding unnecessary costs. Programmatic sampling can also increase research quality in other areas by enabling researchers to collect much more data than through conventional methods. This enables a keener understanding of the insights gathered and can open up a whole new world of opportunities for statistical analyses! How does data science come into play here? Irati: Finding the right model to optimise the attribution logic underlying our programmatic sampling platform is quite complex as we deal with lots of different variables. In our aim to maximise response quality, ensure appropriate targeting, optimise yield per user session, understand and predict user behaviour, and maximise survey completion rates, we have multiple criteria that we need to consider in order achieve top performance. In order for our models to make the right decisions in a matter of milliseconds, we’ve implemented different machine learning algorithms. What exciting projects are you working on? Irati: One of the most exciting projects for me is the trust score: we’re building models that can predict how likely it is that a respondent’s answer accurately reflects reality. For example, if a person says that he/she knows a specific brand, how do we know whether this is true? Building AI models to understand this becomes complex very quickly, but it’s also super-exciting. Another project that we have in our pipeline involves the use of bayesian networks to predict the likelihood that users share certain attributes, for example how likely they are to change their bank account within the next 6 months, without having to ask them. Svitlana: In addition to ensuring that the system always runs smoothly and reliably in a high-growth context, we’re also experimenting with new pricing and yield optimisation algorithms. A recurring area of work is to make sure that our expanding architecture works well with other business lines and requirements, so we need to coordinate very closely with lots of different teams in the company. Seems like a lot of work, what are your next hires? Svitlana: Over the coming months, we’re looking to hire another 15 to 20 people at Dalia. For the programmatic sampling space, the position of Technical Product Manager is particularly important as it involves digging deep into the technical performance of our API integrations with clients and becoming the expert on everything from real-time attribution to yield optimisation. We’re looking for someone who is very passionate about technologies but who also has a sense for creating business value and solving intricate problems. Irati: In our team, we’re also looking for a Senior Quantitative Researcher and in the near future a business intelligence analyst. The job-ad for the business intelligence role is not yet published, but if you’re interested, you can always send a message to hr@daliaresearch.com so that we add you to our list of prospects and reach out to you once this vacancy is live. Svitlana and Irati, thanks so much for sharing your insights on programmatic sampling!
An inside look at Programmatic Sampling
56
an-inside-look-at-programmatic-sampling-1398d5b8674e
2018-08-28
2018-08-28 14:00:28
https://medium.com/s/story/an-inside-look-at-programmatic-sampling-1398d5b8674e
false
853
Founded in 2013, Dalia is a Berlin-based technology startup that distributes millions of surveys in over 90 countries to provide research agencies, academia, public institutions, brands and other organizations access to high-quality market & opinion data.
null
DaliaResearch
null
@daliaresearch
contact@daliaresearch.com
daliaresearch
DATA,BRANDS,GLOBAL,MARKET RESEARCH,TECH
DaliaResearch
Data Science
data-science
Data Science
33,617
Dalia Research
Dalia uses mobile technology to feel the pulse of people around the world.
9055dcd25c8b
DaliaResearch
116
52
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-12-25
2017-12-25 20:05:59
2017-12-25
2017-12-25 20:16:11
1
false
en
2017-12-25
2017-12-25 20:16:11
5
139c296287e5
2.237736
0
0
0
BigDataGuys — Blockchain Use Case on creating Proof of existence for digital assets.
5
PROOF OF EXISTENCE FOR DIGITAL ASSETS BigDataGuys — Blockchain Use Case on creating Proof of existence for digital assets. PROOF OF EXISTENCE FOR DIGITAL ASSETS — Blockchain | BigDataGuys The proof of the existence of anything can be hashed. So, anything with a single digital representation can be hashed and stored in the blockchain. Indeed, any user (node) can query whether the element was hashed and added to the blockchain. Furthermore, there are many more examples of things that can be implemented with Ethereum blockchain platform. A Simple Login System using Ethereum In Ethereum, the addresses are (by definition) systems to prove ownership. The rightful owner of the address is who can perform operations with this address. So, this is the consequence of the underlying public-key infrastructure that is used to verify transactions. For example, we can create a login system based on Ethereum addresses. In a login system, it is always created a unique identity that can be managed by whoever can pass a method to prove that the same entity that created the account in the first place is the same entity doing operations now. Actually, most systems have a username and a password. So, anytime the system requires proof of its identity, then it can request the password for that username. But, in Ethereum we already have a system for proving identities, which are the public and private keys. For example, a simple contract (which can be used by any user to validate his ownership of an address) would have the next logic: First, a user accesses a website to login. Then, when the user is not logged in, then the website requests the user to enter his digital address. Then, the backend for the website receives the address for the user and creates a “challenge string” (to verify his identity) and a JSON Web Token (JWT). Then, these are sent back to the user. Then, the user sends the “challenge string” (used to verify his identity) to the “Login contract” and stores the JSON Web Token (JWT) for later use locally. Then, the backend listens for login attempts using the “challenge string” (used to verify the user’s identity) at the Ethereum network. When an attempt with the “challenge string” for the right user is seen, then it can assume that the user has proven its identity. Furthermore, the only user that can send a message with a digital address is the holder of the private key, and the only user that knows the “challenge string” is the user that received the challenge through the login website. Furthermore, the user gets notified or polls the website backend for confirmation of his or her successful login. Then, the user must use the JSON Web Token (JWT) issued in step 2 for accessing the website. Or, a new JSON Web Token (JWT) can be issued after a successful login. Want to learn more on how to write Ethereum smart contract to validate his ownership of an address? Consult us or get trained through one of our Blockchain Bootcamps at BigDataGuys Main Office: 1250 Connecticut Ave NW, Washington, D.C 20036 Phone: 202–897–1944 | 202–897–1966 For Training — Training@Bigdataguys.com For Consulting- Info@Bigdataguys.com Twitter https://twitter.com/BigDataGuys Linkedin https://www.linkedin.com/company/13190381/ About Me https://about.me/bigdataguys www.bigdataguys.com
PROOF OF EXISTENCE FOR DIGITAL ASSETS
0
proof-of-existence-for-digital-assets-139c296287e5
2018-05-19
2018-05-19 06:36:03
https://medium.com/s/story/proof-of-existence-for-digital-assets-139c296287e5
false
540
null
null
null
null
null
null
null
null
null
Blockchain Technology
blockchain-technology
Blockchain Technology
13,452
BigDataGuys https://www.bigdataguys.com/ @Medium
Artificial Intelligence, Deep learning, Blockchain Bootcamps, Workshops, Platforms and Consulting. Call 202–446–1670, Enroll today!
20dc99cb93f4
TheBigDataGuys
9
3
20,181,104
null
null
null
null
null
null
0
null
0
1c1378e1d03a
2018-07-10
2018-07-10 14:04:10
2018-07-10
2018-07-10 14:09:58
3
false
en
2018-07-10
2018-07-10 14:09:58
4
139dcf028315
5.75
0
0
0
Short version of this post: If you plan to send PIIs (Personally Identifiable Information — information such as social security numbers…
3
Clarifying some important practical concepts about cryptography Short version of this post: If you plan to send PIIs (Personally Identifiable Information — information such as social security numbers, emails, phone numbers, etc.) to a web analytics tool, you should at least use the SHA256 hash algorithm, preferably with some salt. In addition, always ensure the same encoding in all data submissions (Base64 / Hexadecimal, Uppercase / Lowercase) to facilitate connection. The discipline of cryptography has the unfortunate property of having practical applications for virtually everyone in the current digital age and is so vast and complex that it would be unreasonable to expect all these people to have a good handle on it. As a results, some important concepts are misunderstood or simplified in a dangerous way. The objective of this post is to clarify some of these concepts, focusing mainly on the world of web analytics, but we believe that the information can be of more general use as well. In the following we will outline the definitions and uses of symmetric/asymmetric key algorithms, hashing algorithms and encodings, in order to make it clear that they are all different concepts. While not everyone needs to be a security expert, those involved in the more technical processes should have a clear idea of the distinctions. Initial definitions and hashing algorithms There are three types of encryption algorithm: symmetric key, asymmetric key, and hash functions. An extremely important feature separating the uses of the first two from the latter is the reversibility of the process. In the case of symmetric or asymmetric key algorithms, the primary goal is to produce messages whose original content can be retrieved by the recipient, without others being able to read or compromise it. Therefore, it is important that the process is reversible by means of a key. The idea that it is not possible to reverse hash functions to obtain the original message may sound scary at first, seeing as how this is the correct method of sending PII-based identifiers to Analytics. However, for most practical applications, this is not a problem. Let’s suppose you already use a social security number in your business system and want to send it to Google Analytics as a User ID or key to offline data intake. As it is PII, you are not allowed to send it to Google Analytics without using a hash function (at least SHA256 with use of salt recommended, according to this article). Let’s look at the possible uses you might have for this: Possible — Use as unique identifier (User ID): The use of a good hash function virtually guarantees that two distinct social security numbers will generate two distinct codes. That is, just as each user is represented by a unique social security number on your system, they will be represented by an equivalent unique code in Google Analytics. Possible — Import user data from your system into Google Analytics: Suppose you have some information available about the user with Social Security Number A in your database, such as sex or age. The use of a hash function ensures that the application of this on Social Security Number A always returns the same code. So, just apply the same algorithm implemented in online data collection to the file that will be sent to Google Analytics, and the connection will occur without problems. Possible — Extract Google Analytics navigation data to your system: This item is a little more complicated than the previous one, but it follows a simple idea: It is not forbidden for the client to maintain a relation between social security numbers and codes generated by the hash function in their system. This way, by exporting the navigation data crossed with the Google Analytics codes, you can cross-reference it with your internal base. Not Possible — Enrich standard Analytics reports with PIIs: It is possible to send data (the email of a user, for example) after passing it through a hash function, and later, to assemble a customized report with that information. In the meantime, the report would not be legible, as it would only present the final codes. It would also not be possible to search for people by their names in the tool’s standard reports if you need to analyze the behavior of a specific individual. About encodings This topic is special because encoding is not the result of a cryptographic algorithm, but we realize that there are two common sources of confusion and we would like to clarify them. Technically speaking, encoding consists of mapping the byte representation of computer information as symbols, such as letters or numbers. For example, a common internet encoding system is UTF-8, translates byte strings into Arabic or Chinese characters or numbers that can be displayed to users. Other encodings, with different communication uses, are hexadecimal, Base64, URL encoding (used in HTML requests), etc. The first major source of confusion arises from the fact that the result of cryptographic algorithms is usually a string of bytes. Yes, we could display a sequence such as 01110111 on screen… but that would take up a lot of space. Therefore, it is common for coding such as Base64 or hexadecimal to be used to represent such information i.e. in addition to the cryptographic algorithm, it is necessary to choose a coding scheme. Normally it is not difficult to convert from one representation to another, but in some contexts like importing data from Google Analytics it is important to choose only one and use the same for all methods of sending information. The second source of confusion is a direct consequence of this ease of conversion: Encoding provides no security or legal guarantee whatsoever, because anyone can convert it (i.e. it is easily reversible). In Javascript, the atob and encodeURIComponent functions both generate encodings (Base64 and URL encoding, respectively), so they should not be used to mask PIIs when sending to Google Analytics, for example. To be clear about the two points mentioned above, I suggest that you explore the following online tool, which generates hashes using several different SHA methods and provides hexadecimal and Base64 representation: https://approsto.com/sha-generator/ Returning to Reversible Algorithms Normally, it is not common to work directly with symmetric or asymmetric key cryptography algorithms, so almost all direct uses of these techniques are most likely wrong. This may seem strange, because obviously all network communication and even some stored data (e.g. Google Drive, BigQuery) are encrypted. However, these are not simple applications of algorithms in isolated contexts. They are complex protocols that combine multiple different techniques. You can find a good analogy in this article: “AES is to secure cryptosystems what uranium oxide pellets are to a working nuclear reactor.” In this case, AES is a symmetric key algorithm specification. We like this because it explicitly states how dangerous the use of these technologies can be. The reality is that the best in the world make mistakes when implementing these protocols. Although there are risks in using hash algorithms, the most common applications are simple enough for it not to be a problem. In addition, as they are irreversible by nature, the security guarantee is already a little higher. In conclusion, a table Despite the great technical differences between these concepts, in this post we differentiate them using three main characteristics: Is it reversible? This defines whether it is possible to retrieve the original information. Again, several practical applications do not require this to be possible, so a non-reversible method can still be useful. Does it protect information? If it is possible to retrieve the original information, how difficult is it? Using methods that do not provide any protection is a common mistake and can be disastrous. Is it commonly used? How common is the direct use of these techniques in everyday life? The unnecessary use of more sophisticated techniques can be dangerous. We hope that the distinction between these three concepts is clearer after reading, especially in the case of using hash algorithms, clarifying why they are recommended for sending PIIs. Profile of the author: Gabriel Higa Takeshi | Graduated in Computer Science from USP, he currently works as a data engineer and analyst at DP6. Main areas of interest: Statistics, optimization and machine learning.
Clarifying some important practical concepts about cryptography
0
clarifying-some-important-practical-concepts-about-cryptography-139dcf028315
2018-07-10
2018-07-10 14:09:59
https://medium.com/s/story/clarifying-some-important-practical-concepts-about-cryptography-139dcf028315
false
1,378
Blog Posts made by the Digital Marketing Consultancy DP6
null
null
null
DP6 US Blog
info@dp6.com.br
dp6-us-blog
ANALYTICS,DIGITAL MARKETING,MOBILE MARKETING
null
Security
security
Security
41,279
Blog DP6
null
3794c7b4a281
dp6blog
84
7
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-10-06
2017-10-06 04:15:40
2017-10-18
2017-10-18 04:31:00
1
false
th
2018-01-10
2018-01-10 03:44:35
3
139e8bde7014
0.509434
0
0
0
เรามักจะเห็นข่าว หรือภาพยนตร์ พูดถึง A.I. อยู่บ่อยๆ แต่มีใครรู้บ้างว่าจริงๆแล้วมันคืออะไร?
5
อะไรคือสิ่งที่เรียกว่า “A.I.” http://www.freepik.com/index.php?goto=74&idfoto=1086720 เรามักจะเห็นข่าว หรือภาพยนตร์ พูดถึง A.I. อยู่บ่อยๆ แต่มีใครรู้บ้างว่าจริงๆแล้วมันคืออะไร? A.I. ย่อมาจาก Artificial intelligence หรือ ระบบปัญญาประดิษฐ์ มันคือการทำให้ระบบคอมพิวเตอร์มีความสามารถคล้ายคน จนถึงขั้นคิดเองได้เลยทีเดียว ในปัจจุบันมีการนำ A.I เข้ามาใช้งานแล้วในหลากหลายรูปแบบ ที่ใกล้ตัวเราที่สุดก็คือ “สิริ” (Siri) บนไอโฟน นั่นเอง สิริ เป็นผู้ช่วยส่วนตัว ที่สามารถจัดการให้คุณได้ทุกเรื่อง ไม่ว่าจะส่งข้อความ โทรออก เชคปฏิทิน หรือจองร้านอาหาร Robo-Advisor ก็เป็นอีกหนึ่ง A.I. ที่กำลังมาแรงในสาย Fin Tech คือการใช้ A.I. ในการวิเคราะห์ และจัดพอร์ตให้ลูกค้า (ง่ายๆ คือเล่นหุ้นแทนเรานั่นเอง) ที่ฮิตที่สุดตอนนี้ คงหนีไม่พ้น Chat Bot เพราะเป็น A.I. ที่เริ่มมาแทนที่คน อย่าง Chat Bot ของสำนักข่าว CNN เมื่อเราส่งหัวข้อข่าวที่เราสนใจไป บอทก็จะทำการสรุปข่าวเกี่ยวกับเรื่องนั้นๆมาให้ โดยบอทจะทำการเรียนรู้สิ่งที่เราถามไปเรื่อยๆ เพื่อที่จะส่งข่าวให้ตรงกับความสนใจของเรา เชื่อว่าในอนาคต A.I. จะถูกเอามาใช้งานแทนคนมากขึ้น เราจึงต้องพร้อมที่จะเรียนรู้ และปรับตัว เพื่อที่จะใช้ A.I. ให้ได้ประโยชน์สูงสุดนั่นเอง 📢 http://www.trainingcenter.co.th/ ☑️Facebook Page: https://www.facebook.com/ntcfanclub ✅ LINE Official: @NTC-LINE ☎️ Tel: 0–2634–7993–4 Ext.11–17
อะไรคือสิ่งที่เรียกว่า “A.I.”
0
ai-139e8bde7014
2018-01-10
2018-01-10 03:44:36
https://medium.com/s/story/ai-139e8bde7014
false
82
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Network Training Center (NTC)
null
aac6a08bdff2
networktraining
18
17
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-07-08
2018-07-08 14:25:21
2018-07-08
2018-07-08 14:36:54
0
false
en
2018-07-09
2018-07-09 00:58:17
2
13a004236fd0
2.4
17
3
0
Let’s just admit we’re all fighting for ourselves.
5
I hate Diversity & Inclusion… Let’s just admit we’re all fighting for ourselves. I hate diversity and inclusion because I feel it cheapens my worth. I know I’m the best at what I do. Ask any client of mine. Not because I have to do it 10x better than some trust fund baby just to get recognized as existing, but because at my craft, I have put in the time, effort and discipline to become a globally recognized master of the art. I would be a master even if I didn’t have to compete with the legacy of benchwarmers who have a 400 year head start. I am what I am not because of diversity and inclusion but because of my mom. This 5’2 and shrinking Mexican who grew up in as a diverse and inclusive environment as 1950s Los Angeles tended to be for Chicanos. You know, before Watts had yoga studios and those real ganstas in blue, the LAPD, got diversity training and non-lethals. I also hate diversity and inclusion because when we have these conversations, WE have them amongst OURSELVES. Why are we all talking to each other? It’s like group therapy where we all lament the trials and tribulations. Reliving each police killing shot by shot, each kid stolen from a screaming parent tear by tear. My #metoo started in 1619 and as a society we haven’t cared since then. As at the end of the day, we’re only fighting for ourselves, our families and what we believe to be true. There are the times though, where I appreciate diversity and inclusion. Cause the person goes from asking me to get them coffee, valet their car, or “If they can help me” through that privileged veneer that translates in tone to “Nigger, WTF are you doing here?” — to realizing just a moment too late — they should have always been in fear of me. Not because I’m Black, Mexican and from Texas. But because I’m there to take their customers, clients, income and very livelihood. Because wherever “there” is, from DC to Dubai, to Dublin, I am there fighting for myself, my family and my legacy. So I do appreciate the tactical advantage of being underestimated by many through the D&I lense. If you hate diversity and inclusion like I do, come join us on Tuesday, July 10 2018, 6PM EST at Bowie State University as we discuss Diversity, Inclusion, Blockchain and The Road Ahead. Cause while I hate diversity and inclusion, I understand, why as a whole, we must discuss it, debate it and ultimately seek to make the conversation irrelevant. But until we get pass the historical legacies of institutional racism and being in denial about the continues effects of slavery on our psyche and society — bring your isms and join us for a frank conversation on this chapter of Digital Transformation, known as Blockchain. Why? Because the chapter after this is called “The rise of AI”. And the machines don’t care what socially construed version of a “race” you are. They just want to automate your job and make you as irrelevant as “diversity & inclusion” in an automated ecosystem. Ps — It’s a safe place. I won’t judge you or your -isms. All are welcome to join. — — — — — — — — — — — — — — — — — — — — — — — — — — — — — — My name is Samson Williams. I’m a human and an anthropologist. If you agree, like it. If you disagree, please comment and tell me why. Feel free to hit me up on Twitter or Instagram @HustleFundBaby or connect with me on LinkedIn. Finally, I would say thoughts are my own but I probably stole them from a woman.
I hate Diversity & Inclusion…
88
i-hate-diversity-inclusion-13a004236fd0
2018-07-09
2018-07-09 00:58:17
https://medium.com/s/story/i-hate-diversity-inclusion-13a004236fd0
false
636
null
null
null
null
null
null
null
null
null
Diversity
diversity
Diversity
17,812
Samson
Cheerleader of all things startup and entrepreneur. Life's a hustle, invest in something worthy of you. @AxesAndEggs @UNHLaw #Blockchain #Cryptocurrencies
7177725d7720
HustleFundBaby
186
218
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-09-17
2018-09-17 02:43:19
2018-09-17
2018-09-17 02:55:25
1
false
en
2018-09-17
2018-09-17 02:55:59
10
13a0b38b04e5
2.037736
0
0
0
Deep learning has achieved a big success in the past few years, but its interpretive power is limited. They work largely because of the…
2
Use of Graph Networks in Machine Learning Deep learning has achieved a big success in the past few years, but its interpretive power is limited. They work largely because of the abundance of data. On the other hand, traditional machine learning algorithms are much better in interpretive power, but manual feature engineering costs a lot, due to the lack of data in earlier era. In light of this, a group of scientists initiated the work of graph networks, aiming at devising new artificial intelligence algorithms that exploits the advantages of two worlds, while still holding the principle of combinatorial generalization in constructing methods by using known building blocks to build new methods. Graph is good at interpretation as it is good for relational representation. The use of graph networks is more than the graph convolutional neural networks (GCN) in the previous two blog entries. (part I and part II) However, to achieve relational inductive biases, an entity (an element with attributes), a relation, (a property between entities) and a rule. (a function that maps entities and relations to other entities and relations) This can be realized using graph, which is a mathematical structure that contains nodes and edges (that connect nodes.) To generalize the use of graph networks in various machine learning and deep learning methods, they reviewed the graph block, which is basically a function, or a mapping, from a graph to another graph, as shown in the algorithm below: Works of graph networks are not non-existent; the authors listed previous works that can be seen as graph networks, for example: Message-passing neural network (MPNN) (2017); Non-local neural networks (NLNN) (2018). The use of graph networks, I believe, is the next trend. There have been works regarding the graph-powered machine learning. (see Google AI blog, GraphAware Slideshare) I recently started an open-source project, a Python package called graphflow, to explore various algorithms using graphs, including PageRank, HITS, resistance, and non-linear resistance. Peter W. Battaglia, Jessica B. Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinicius Zambaldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, Ryan Faulkner, Caglar Gulcehre, Francis Song, Andrew Ballard, Justin Gilmer, George Dahl, Ashish Vaswani, Kelsey Allen, Charles Nash, Victoria Langston, Chris Dyer, Nicolas Heess, Daan Wierstra, Pushmeet Kohli, Matt Botvinick, Oriol Vinyals, Yujia Li, Razvan Pascanu, “Relational inductive biases, deep learning, and graph networks,” arXiv:1806.01261 (2018). [arXiv] “Graph Convolutional Neural Network (Part I),” Everything About Data Analytics, WordPress (2018) [WordPress] “Graph Convolutional Neural Network (Part II),” Everything About Data Analytics, WordPress (2018) [WordPress] Sujith Ravi, “Graph-powered Machine Learning at Google,” Google AI Research Blog (2016). [Google] Sujith Ravi, Qiming Diao, “Large Scale Distributed Semi-Supervised Learning Using Streaming Approximation,” arXiv:1512.01752. (2015) [arXiv] Vlasta Kus, “Graph-Powered Machine Learning,” GraphAware. [SlideShare] Maksim Tsvetovat, Alexander Kouznetsov, Social Network Analysis for Startups, O’Reilly (2011). [O’Reilly] PyPI: graphflow [PyPI]; Github: stephenhky/GraphFlow [Github] Original article: https://datawarrior.wordpress.com/2018/09/16/use-of-graph-networks-in-machine-learning/
Use of Graph Networks in Machine Learning
0
use-of-graph-networks-in-machine-learning-13a0b38b04e5
2018-09-17
2018-09-17 02:55:59
https://medium.com/s/story/use-of-graph-networks-in-machine-learning-13a0b38b04e5
false
487
null
null
null
null
null
null
null
null
null
Machine Learning
machine-learning
Machine Learning
51,320
Stephen Ho
Applied Quantitative Researcher
ca4dae1bca42
stephenhky
105
205
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-06-07
2018-06-07 18:50:39
2018-06-18
2018-06-18 15:02:03
2
false
es
2018-06-18
2018-06-18 21:59:30
1
13a19321ded7
4.802201
8
0
0
Por: Pamela Armstrong consultora UX senior y docente, Consuelo Pizarro C. consultora UX/UI senior.
5
Diseñando una experiencia digital de salud para pacientes/usuarios. Parte 1 Por: Pamela Armstrong consultora UX senior y docente, Consuelo Pizarro C. consultora UX/UI senior. En este artículo relataremos los aprendizajes de nuestra experiencia como equipo de UX-UI diseñando un portal de salud que concentra la información de clínicas, centros médicos y clínicas dentales. El desafío era homologar en un solo lugar información que cada negocio trataba de manera distinta, considerando a nuestro usuario/paciente en primer lugar y su necesidad de resolver un problema de salud. Por otro lado debíamos contemplar necesidades del negocio y del área médica. Trabajamos sobre un modelo que integra los sitios web de las clínicas y centros médicos, buscando optimizar la operación en red para la organización y los clientes. En esta integración debíamos considerar: 9 clínicas, 3 clínicas dentales y 32 centros médicos. El encargo que recibimos como equipo era diseñar una plataforma que incluyera a todos los prestadores de salud juntos. Para eso, desde ese momento hacia el futuro, la idea es generar un HIS (Hospital Information System) que concentrará toda la información que se comience a reunir desde el sitio web, la agenda y en el futuro las fichas de los pacientes y funcionamiento interno. Nuestro objetivo principal era Diseñar un portal que permitiera: Encontrar información sobre clínicas, clínicas dentales y centros médicos. Agendar desde el portal, con cualquiera de los prestadores/médicos. Teníamos que incluir la información y funcionalidades de todos los sitios web de las clínicas y centros y disponibilizar la información de cada uno y también sus diferencias de negocio, sin perder el mayor atributo de cada una, por ejemplo la clínica X tiene como foco la maternidad y todos los servicios asociados a este tema, la clínica Y se enfoca en segmentos medios-bajos de clientes y tiene convenios con Fonasa. Había que generar una nueva identidad, ahora serían todas las clínicas y centros juntos una sola cosa y desde este momento la operatividad del negocio sería la misma para todos.. Teníamos la oportunidad de diseñar algo nuevo, más cercano y útil para los usuarios/pacientes. Los 12 aprendizajes A medida que fuimos realizando el proyecto con el equipo, nos enfrentamos a diversas situaciones, de las que aprendimos mucho sobre aspectos que no están totalmente mapeados como parte del rol de UX como tal, pero que, de cierta forma, son parte de la gestión de un proyecto exitoso. Nuestros aprendizajes fueron 12. Generar equipo y un espacio de confianza Desde el primer momento tuvimos claro que la relación con el cliente sería clave. Era un proyecto de varios meses de trabajo y necesitábamos generar una buena relación que nos permitiera hablar con confianza y no tener miedo de contradecir a nuestra contraparte sin caer en enfrentamientos y conflictos. Realizamos talleres, jornadas de trabajo y revisión, y reuniones semanales (donde la comida era un detalle importante). Además cada miembro del equipo, desde su área de expertise, era capaz de comunicarse con el cliente para no depender exclusivamente de una persona en particular. Gracias a lo todo lo anterior, logramos que el cliente se sintiera parte de nuestro equipo. Lo empoderamos para que fuese capaz de defender con argumentos claros las decisiones de diseño, como si fuese uno más de nosotros. 2. No llevar problemas, ofrecer soluciones Sabíamos que el cliente no tenía dedicación exclusiva para este proyecto, por lo que cada vez que detectamos un problema, pensamos posibles soluciones. La idea era detectar el problema pero siempre ofrecer alternativas de solución. Uno de los principales problemas que descubrimos fue que cada clínica hablaba un lenguaje distinto, por lo que generamos sesiones de trabajo y entrevistas para entender mejor la información. Propiciamos la comunicación entre los diferentes stakeholders y generamos matrices y definiciones de servicios, unidades, programas y centros. La tarea más compleja fue la homologación de los servicios y las especialidades, para lo cual se realizaron comisiones y se estableció un calendario de reuniones con el fin de tener fechas de entrega claras. 3. Adelantarse a los problemas, coordinar y conducir Detectar el problema era el primer paso, el siguiente era orquestar todos los factores que llevasen a buen puerto nuestras propuestas. Mandar emails, coordinar reuniones, llamar a quien fuese necesario e impulsar la comunicación entre los interesados. El proceso de homologación de servicios y especialidades fue el mejor ejemplo de esto. En este proyecto fue clave nuestra intervención. Tuvimos que impulsar la formación de una comisión médica para la homologación de servicios, unidades y especialidades. Nuestro rol fue moderar estas instancias y hacer que los integrantes de la comisión pensaran en el lenguaje más cercano al usuario. Para apoyar la encontrabilidad, diseñamos un motor de búsqueda predictiva, que estaría conectado con la información disponible en el sitio, con artículos y estrategia de SEO, de tal forma que el usuario pueda llegar a la información precisa tanto desde dentro como fuera del sitio. 4. Tener el rol de facilitador El estar a cargo del diseño de un nuevo portal de salud fue una tarea difícil y compleja, considerando que debíamos unificar necesidades del negocio, comerciales, de la nueva marca, del área médica y de los pacientes/usuarios. Por lo tanto, nos propusimos hablar con todos y atender los requerimientos de cada área. Realizamos talleres de estrategia con los representantes de cada clínica; reuniones con el cliente; reuniones con la agencia de branding; sesiones de trabajo con médicos para la homologación, confección de prototipos navegables; desarrollo de un webkit y documentación de cara a la etapa de desarrollo. 5. Mantener un norte claro en todo momento Cada vez que nos sentimos perdidos o tuvimos dudas, recurrimos a los pilares de marca y a la estrategia de la experiencia digital completa, los que estaban definidos desde el principio del proyecto. Las decisiones de arquitectura de información se basaron en este ciclo de la experiencia. Nos propusimos priorizar los pilares considerando la persona, la necesidad, el canal, contexto y escenario de uso. Y nos dimos cuenta que, si bien el sitio web debía hacerse cargo de los pilares de marca, el pilar de Disponibilidad era el que debía predominar. ¿Cuál es la motivación del usuario? ¿Qué necesita? Necesita solucionar su problema de salud, reservando una hora de atención lo antes posible, lo más cerca de su lugar de trabajo o residencia. Y nos encargamos de recordárselo al cliente cada vez que sentíamos que nos alejábamos de ese objetivo. 6. No traspasar al usuario el lenguaje del negocio Esto era clave. Lo que es lógico para el negocio no tiene por qué ser lógico para el cliente, y lo que es lógico para los médicos no tiene por qué ser lógico para el paciente. Para el negocio, los exámenes, toma de muestras, imagenología y radiología son mundos distintos; para el usuario son exámenes. Para los médicos, las especialidades son claras, pero los usuarios ¿saben con qué especialista reservar hora según los síntomas que presentan? La mayoría de las veces no. Continúa leyendo el resto de nuestros aprendizajes en la publicación de mi co-autora Consuelo Pizarro C.
Diseñando una experiencia digital de salud para pacientes/usuarios. Parte 1
34
diseñando-una-experiencia-digital-de-salud-para-pacientes-usuarios-parte-1-13a19321ded7
2018-06-18
2018-06-18 23:43:54
https://medium.com/s/story/diseñando-una-experiencia-digital-de-salud-para-pacientes-usuarios-parte-1-13a19321ded7
false
1,171
null
null
null
null
null
null
null
null
null
Healthdesign
healthdesign
Healthdesign
0
Pamela Armstrong
Consultora UX senior, Docente diseño de interacción UTEM, Diseñadora Industrial, Mentora en +MujeresUX
68fb6ddc5472
pamela.armstrong.herrera
19
19
20,181,104
null
null
null
null
null
null
0
null
0
a599fd119b1b
2018-01-26
2018-01-26 13:32:12
2018-01-26
2018-01-26 13:38:37
1
false
en
2018-01-26
2018-01-26 13:38:37
1
13a1d6b2246e
1.532075
1
0
0
Karlsruhe, Germany: 26 January, 2018 — Zana, an interactive Artificial Intelligence-based health assistant was selected as an award winner…
5
ZANA is award winner of INNOLABS Programme for Innovative Health Projects ZANA A.I Karlsruhe, Germany: 26 January, 2018 — Zana, an interactive Artificial Intelligence-based health assistant was selected as an award winner by INNOLABS 1st call for Innovative Health Projects. INNOLABS Programme aims to unlock the cross-sectoral collaborative potential of SMEs by combining ICT with the BIO, Health, and Medicine sectors to deliver market sensitive disruptive technologies. In focus is mHealth as a catalyst and the application of results to personalized health for elderly populations. INNOLABS is funded by European Union’s Horizon 2020 research and innovation programme. Zana is a virtual assistant with medical knowledge. Zana can understand and process in real-time through natural language dialog the information need of the user. Through intelligent recommendations the assistant is able to show concise informative articles and products for personal health management. In the scope of INNOLABS project, Zana partners with Ticino Health SA Switzerland, a health-focused platform comprising regional, national and international specialists from different areas of medical expertise. The collaboration of Zana and Ticino Health would form Europe’s first AI-based solution that supports (elderly) patients with personalized, reliable health information. Zana advises information seekers on possible treatment options and recommends products for their healthy living practice. Ticino Health further provides patients with instant medical advice and treatment offers. This is the second grant awarded to Zana in a short period of time, after having received the prestigious EXIST funding by the German Federal Ministry for Economic Affairs and Energy to support transfer of research technologies into the market. Zana AI is available for conversation on Facebook Messenger, with plans for adoption to other popular messenger platforms such as Skype, Google Allo and amazon Alexa. Health searches attract more than 9 billion queries per month, but the trend is pushing toward using conversation-based platforms, such as Zana, to get accurate information on medical conditions, prevention and treatment options. Zana benefits from a strong team of researchers and software engineers, and a broad network of doctors and clinics. Original Article: ZANA is award winner of INNOLABS Programme for Innovative Health Projects
ZANA is award winner of INNOLABS Programme for Innovative Health Projects
1
zana-is-award-winner-of-innolabs-programme-for-innovative-health-projects-13a1d6b2246e
2018-05-03
2018-05-03 16:38:17
https://medium.com/s/story/zana-is-award-winner-of-innolabs-programme-for-innovative-health-projects-13a1d6b2246e
false
353
Zana is an intelligent assistant that responds to health questions with trusted medical information
null
ZanaAlpha
null
Zana AI
info@zana.com
zana-ai
AI,HEALTHCARE,MACHINE LEARNING,CHATBOTS,NATURAL LANGUAGE PROCESS
zana_assistant
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Armand Brahaj
Researcher, Software Developer, Business-Runner, Activist.
72136e515418
alpukn
5
7
20,181,104
null
null
null
null
null
null
0
null
0
8a5c79a9c9e6
2017-11-08
2017-11-08 11:10:32
2017-11-08
2017-11-08 11:33:03
4
false
en
2017-11-08
2017-11-08 13:24:47
1
13a1ee019bbe
2.281132
10
0
0
Congratulations to the BotSupply team✨
5
Danish AI startup BotSupply wins the IBM award 2017 Congratulations to the BotSupply team✨ In occasion of the Watson Summit Denmark 2017, BotSupply won the IBM Award 2017. The event took place at the stunning facilities of DR Koncerthuset where AI Enthusiasts and experts gathered for a day of inspiring knowledge sharing. The main focus of the summit was allocated to cognitive technologies and AI. According to IBM Research Cognitive technologies surpass all other technologies because of their ability to understand, see, hear and learn. Within a few years, businesses will be ‘cognitive at the core’. This in will in turn strengthen the foundation of the business and the decisions that are being made on a continuous basis. Watson Summit was targeted to decision makers and leaders in strategy, operation, finance, HR, IT, marketing, customers and sales. Attendees were presented with the latest knowledge on cognitive technology and concrete examples of new business opportunities in banking and insurance, health, industry, retail and government. Asser Traberg Smidt one of the founders of BotSupply receiving the IBM Award 2017 As part of the summit BotSupply was chosen along with 5 other startups that work with: AI, Cognitive Computing and Big Data Analytics to participate in the Startup Event. The Startup Event consisted of a competition framed around the question: “Who would you invest in if you had 1 million DKK?” The 6 competing startups presented a pitch that highlighted how they respectively use technology to create maximum value, how their startup is built up and explained their business model. A sharp and talented panel of juries made by experts in the field and investors chose unanimously BotSupply as the Startup that showed the most promise and potential, awarding BotSupply with the IBM Award 2017. The prize consists of a trophy and a check for 120.000 USD worth of Bluemix credits to be used on Watson technologies. The BotSupply team is all smiles! The BotSupply team looks forward to persisting in dedicating all its efforts on continuing to help digitize Denmark by creating a new ecosystem for AI. This new ecosystem being powered by AI Scientists, Bot Engineers and Creatives that leverage the capacity of cutting edge technologies for creating top notch cognitive solutions for companies searching to implement AI. We call this process Co- Creation of AI. By co-creating AI, BotSupply seeks to lower the entry barrier that companies encounter when looking to implement AI, when they find themselves unable to find the talent that is capable and able of developing tailored cognitive solutions. You can learn more about the exciting projects that Botsupply is currently working on at: www.botsupply.ai
Danish AI startup BotSupply wins the IBM award 2017
181
danish-ai-startup-botsupply-wins-the-ibm-award-2017-13a1ee019bbe
2018-05-11
2018-05-11 14:32:16
https://medium.com/s/story/danish-ai-startup-botsupply-wins-the-ibm-award-2017-13a1ee019bbe
false
419
Organizations partner with our network of AI scientists, bot engineers and creatives to co-create AI & bots
null
BotSupply
null
#WeCoCreate
yourfriends@botsupply.ai
botsupply
MACHINE LEARNING,ARTIFICIAL INTELLIGENCE,CHATBOTS,DEEP LEARNING,UX DESIGN
botsupplyhq
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Grasia Hald
🤖 AI enthusiast & SoMe junky✌🏽 Sometimes I am a communicator, sometimes I am a Neuropsychologist. I’m always curious about disruption & bots.
b476a0891380
gmhald
86
25
20,181,104
null
null
null
null
null
null
0
null
0
6ce54c7380cf
2017-09-23
2017-09-23 17:13:25
2017-09-23
2017-09-23 18:14:58
2
false
en
2017-09-25
2017-09-25 13:10:58
1
13a230ef0558
1.621069
3
0
0
Cornea AI makes it to the ‘New Apps We Love’ list on the iTunes Store
5
Autumn starts with a BANG! Autumn has never been so amazing. In addition to the cozy nights, beautiful leaves, flickering flames of the camp fire, the iTunes top apps list has made the start of Autumn tremendous and exciting. Cornea AI is now considered as one of the top apps in 2017 that people love to use and experience. Cornea AI featured in the ‘New Apps We Love’ iTunes Store How did Cornea AI reach the top position? Cornea AI is an artificial intelligence powered photo editing app that uses deep learning to predict the popularity of a picture before sharing it on social media. Every single individual doubts the popularity of a picture before sharing it on Instagram, Facebook or any other social networking platform. We all think twice about the number of likes, hearts and impressions we get on our photographs. So why not just sit back and relax? Let Cornea AI predict the popularity of your photos. In just one tap, you can share the most stupendous edits of your photographs. Using Artificial Intelligence, we provide every edit made in Cornea AI with a score. This is your popularity score and is also known as ‘Cornea Score’. The higher your Cornea Score, the higher chance your photograph has to go viral. Have a look at a Cornea Score transformation guide here. Cornea AI also lets its users to select from trending hashtags provided by the ‘Cornea Hashtag Recommendation’ before posting on Instagram. These hashtags fetch you those extra likes and hearts from all around the globe and make a photo go viral. The main aim of Cornea is to democratize creativity, so that sharing a great photo is as easy as composing a new tweet- no fiddling with many editing tools, collage options or filters, let AI figure out what will work best for you and recommend you accordingly. So, in all our busy schedules, lets find ourselves some time to share a Cornea edited photo and celebrate Cornea’s presence in the top apps iTunes list.
Autumn starts with a BANG!
101
autumn-starts-with-a-bang-13a230ef0558
2018-02-13
2018-02-13 19:10:04
https://medium.com/s/story/autumn-starts-with-a-bang-13a230ef0558
false
328
KARNA leverages the power of AI to provide unique solutions by generating valuable insights from unstructured texts and videos to empower and automate market research
blog.karna.ai
karnaAI
null
KARNA AI ( Market Research Division of ParallelDots )
contact@paralleldots.com
karna-ai
ARTIFICIAL INTELLIGENCE,EYE TRACKING,TEXT ANALYSIS,OBJECT DETECTION,IMAGE RECOGNITION
ai_karna
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Cornea App
Using Artificial Intelligence to take the guess work out of photo sharing
f1226e7fedec
cornea_app
8
2
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-10-03
2017-10-03 09:18:30
2017-10-03
2017-10-03 14:59:52
1
false
en
2017-10-03
2017-10-03 15:22:37
71
13a2d1944f5f
9.316981
2
0
0
An 8 min read on how to solve for trust in AI, so that we let intelligent machines create more value than what they will destroy.
5
“What is well thought is clearly explained”, Nicolas Boileau “The Art of Poetry / 1674” , but also attributable to “Explainable AI / 2017”. Decentralised trust: A positive alternative to the AI economic dystopia ? (An 8 minutes read, unless you click on all links and read all background pieces of journalism & research that made this story possible). You most certainly came across the latest tech catchy headlines: AI will take away all sorts of jobs without creating new ones, and deep-learning technology is leading us to lose control on intelligent machines. According to many publications, our society and economy would be doomed… But as I’m wearing a banking hat on engineering brains, I tend to think the economic issue at stake is more about Trust than Technology: Trust is, and has always been, at the core of trade and the economy… and right now, we’re missing some of it. If anything, it’s actually our defiant attitude towards AI, more than AI itself, that presents a threat to the global economy. To find a better future than the one described in many AI dystopia stories, we need to address the following question: How to solve for trust in AI so that we let intelligent machines create more value than what they will destroy? And hopefully, below is one answer. PART I / The four forces behind the AI economic dystopia: Threat, Opacity, Difference and Reputation. AI IS A THREAT: The 4th industrial revolution* predicted 65 years back has happened, and you cannot reverse it. Driverless cars, smart factories, robo-advisors, personal assistants and other intelligent machines are spreading fast. Since its first online book sale 20 years ago**, Amazon.com did move on to become a global provider of artificial intelligence services. Knowledge, our former economic engine, has been commoditized: Centralized systems are now able to deliver instant personalized advice and services to large audiences with quasi-infinite scale, and with greater expertise than isolated human experts. In simple words: You are now actually already living the aftermath of a revolution that has turned the knowledge economy upside down***. Many feel threatened or are tempted to resist the use of AI, and even Elon Musk described AI as humanity’s biggest existential threat that could lead to potential wars. But just as English textiles workers Luddites destroyed weaving machines in the 19th century as they saw their job and social status threatened by a new piece of technology****, neo-luddites are today only uselessly resisting unstoppable exponential changes coming from those virtual brains that we design and train, but that we still often don’t trust. AI IS OPAQUE: From banking to health and military, intelligent machines are now making decisions we cannot explain. So, how to trust them? As we’re progressing on our digital agenda at the bank, I had similar conversations three times on different topics recently: intelligent OCR, machine learning for credit decisions, and predictive analytics for global trade. Each time, the same question was asked: “How can you explain the decision or prediction the machine came up with?”. And each time the same answer: “If you’re talking about deep-learning based system, then essentially you cannot”. Voluntarily thought-provoking answer, but it is true that with deep-learning technology, i.e a technology that let computers program themselves, even MIT professors and AI researchers recognize they can’t explain decisions made by machines they originally designed. At best multiple industries may soon benefit from research led by U.S Department of Defense to build Explainable AI, but the result will still be nothing like an algorithm process chart explaining the step by step path towards a decision. This will narrow the existing trust gap between humans and computers, but we will eventually still have to take a leap of faith as full decision audit trails may never be available to us. AI IS DIFFERENT: Machines aren’t our peers, so we instinctively don’t trust them. No matter how Descartes has been proven wrong by behavioural neurologists, no matter you heard all about the long list of 186 human cognitive biases described on Wikipedia, you most probably still prefer leaving your life in the hands of an unknown human Uber driver, rather than stepping into a car driven by some opaque neural network running on a microchip. And that’s precisely because of our human in-group bias: It doesn’t matter if the neural network drives better than the human driver, humans don’t perceive the machine as part of the “mankind group”, and hence instinctively reject it. IBM Watson can beat doctors at predicting cancer all it wants, and Unilever can prove many times over that AI is less biased than human when it comes to recruitment…. The reality is that even though human decisions are equally opaque and not as good as machines’ decisions, humans favour other humans’ decisions. Ask yourself: Do you really trust that recommendations made by computers are as good as, if not better than, human decisions? Don’t think only GPS, but think health, or even justice. Do you really trust that computers are engineering better architecture, vehicles, articles, music pieces than humans? Until a majority of humans answers yes with confidence to both questions, we will not be building effectively an AI based economy. Familiarity breeds trust, and trust will improve as we are interacting with AI agents, but it may take long before we consider them as peers, despite Facebook efforts to make bots look more human. AI IS PARTNERING WITH THE WRONG CROWD: Global businesses aren’t that trusted anymore. Despite Global Trust Barometers are displaying an all time low level of trust in global businesses, number of global platforms, networks and brands started to act as “trust aggregators” and leverage their brand name and scale to deploy less known AI based services. For instance, Audi or Mercedes-Benz will soon offer a driverless option for their car thanks to their partnership with the lesser known Nvidia. But global corporations’ already challenged reputation is actually raising more concerns than helping AI spread, especially as consumers question the ability of global firms to avoid scaling biases and maintain strict controls on data privacy. Biased decisions made by an AI engine that serves very large audiences automatically impacts a large number of human lives. In addition, as the AI based services will need a feedback loop to learn, users will have to share their data with a handful of large aggregators, rather than with fragmented operators. To sum up this point: think one super-GP holding the health records and family histories of patients of an entire country, rather than just a neighbourhood… and then think what happens when this super-GP gives the wrong diet advice to an entire population rather than just to a small village, or if his patients database is hacked. IN THE END, AI AND THE FOUR FORCES IS THE COCKTAIL FOR DYSTOPIA: When the four forces mentioned above are combined, the perspectives of the AI economy do look bleak. As long as we trust more human decisions than machines’ decisions despite their equal biases, global brands benefit from a trust arbitrage over unknown service providers while channeling the explosion of AI-based services in a countless number of industries and verticals. It creates a bottleneck that engenders an imperfect centralised economy, i.e one that accelerates imbalances, scales biases, creates concentration risks and raises data privacy concerns. In this scenario, this AI trend automates for the few rather than creating for the many, and rapidly destroys an old and more evenly distributed knowledge economy while posing a systemic threat to society. It’s the background for the long string of news article and a number of books, from The Circle to Homo Deus, portraying the end of our society and economy as we know it. But there’s an alternative end to the AI story…. PART II / Alternative perspectives for the AI economy: Certified “domain training” for intelligent machines. One way to create a more evenly distributed and sustainable AI based economy is to set up a framework under which individual intelligent machines can be trusted as much as humans, and more than brands or networks. This can be largely inspired from the pre-AI knowledge economy, which relied heavily on decentralised trust to encourage the exchange of services, much like a peer-to-peer offline marketplace. Every day, we trusted peers we worked with or bought services from, without referring to any global trust aggregators. We trusted their individual experience, and their qualifications (that’s why GP usually frame their diplomas on the wall). If we’re ready to jump the in-group bias fence and consider AI as peers, then we can rebuild a similar framework for the AI economy letting decentralised certifications foster trust. A number of such AI certifications can already be imagined based on various government and research initiatives: An “Explainable AI certification” would, for instance, be the first certification that any intelligent machine has to pass to prove it can explain (at least up to some degree) its decisions. Some consulting firms argued that companies adopting Explainable AI, i.e being certified if such certificate would exist, would gain competitive advantage. Opacity may never be entirely removed, but it can be reduced and evaluated. “Health & Safety certifications”, beyond just explainability, can be derived from the work currently led by the IEEE Standards Association’s “Ethically Aligned Initiative”. The initiative covers considerations for Transparency, Data privacy, Biases, Control, Fail-safe and even for Alignment to user well-being. Its mission is “to ensure every technologist is educated, trained, and empowered to prioritize ethical considerations in the design and development of autonomous and intelligent systems”. But this shouldn’t be only considerations for technologists building AI systems. Systems themselves could well be certified against similar standards to increase trust, just like regardless of the qualifications of the architect and developer who built a house, no one would move in a new home until it is certified meeting basic security standards. The intelligent agents that now run the house and capture daily life private data to assist homeowners with their chores need to be certified as part of the house. In the end, no one wants to see his neighbours going through an HAL 9000 — Space Odyssey types of scenario. “Functional certifications”, we trust unknown individuals as soon as they are “certified”. Often certification doesn’t provide a 100% guarantee that your counterpart will not be incompetent or unbiased, but it does ensure that the training was of quality, and that it was successfully completed and validated. Investment analysts, neighbourhood GPs and car drivers don’t come “set up”. They are trusted economic agents first because of their certifications, qualifications or license. On top of academic degrees, the list of professional certifications is long, and before any experience, word-of-mouth or network effect network kicks in, those certifications are the first seed that establishes trust around individual experts. Today, there is still no equivalent of a CFA Level III or PMP certification for machines… and hence no space of newcomers AI based professional agent to be trusted. Interestingly, if the role of regulators is to preserve general interest, and if machines are proven to preserve general interest better than humans, then in this scenario regulators would have to rule in favour of machines. Just like bicycles are not allowed on the highway, soon would come a time where only driverless cars are allowed on the highway. Similarly, in many industries, professionals and companies operating without certified artificial intelligence assistance may not be allowed to operate. This wouldn’t mean that we are all to be replaced, but it would most probably mean that we would all need to train our assistants and get them certified. AI training still required very large scale businesses, but as technology progresses and intelligent agents are able to learn from smaller data sets, it won’t take long before AI training becomes mainstream. The AI economy would then be unleashed as individuals and SMEs would train agents not only to assist them but to offer new services according to the company-specific know-how under a safe regulatory framework and certification preserving customers’ interests. The resulting decentralisation of Trust would rewrite the next chapter of AI, which today reads as a world led by trust aggregators imperfectly and unilaterally capturing all economic value from AI. * 4th Industrial Revolution: The first book where the expression “the 4th industrial revolution” was ever mentioned was “America’s last chance”, by Albert Carr, and was actually published 65 years ago. It then urged U.S to break with isolationism, and predicted that a revolution underpinned by air transport and radio communication would lead to a concentration of decision power and a globalised world. Taking action was then presented as America’s Last Chance to stay relevant in this upcoming new world. ** Amazon, Books & AI: The first title sold by amazon.com was Fluid Concepts. A 1995 book on Artificial Intelligence written by Douglas Hofstadter, arguably one of the most articulated author on the topic. Coincidentally since then, Amazon expanded from an online bookshop to become a global provider of artificial intelligence & cloud services. In other words, the 4th industrial revolution has happened before your eyes, though you may not have noticed it. Like for many others, there is a good chance you are simply blind to it either because you didn’t grasp the extent of the revolution, or because you fully understand it but some form of unconscious neo-luddite angst is pushing you to ignore the subject. In both cases, a proactive deep look at what’s happening is important for your career in the years to come. *** Knowledge economy & AI: Countless news publications express how, from workers to cashiers, to store managers, to radiologists, to dermatologists, to lawyers, to journalists, to songwriters, to engineers, to taxi drivers, to delivery men, to farmers, to bankers, to apple pickers…. Workers along with knowledge workers are all going to lose their jobs to machines. It is often said, “automation takes the job nobody wants, artificial intelligence takes the job everybody has”. Finally, you know software fear has peaked when it’s actually Bill Gates who suggests taxing robots. **** Luddites: The Luddites were a group of English textile workers and weavers in the 19th century who destroyed weaving machinery as a form of protest. Luddites feared that the time spent learning the skills of their craft would go to waste as machines would replace their role in the industry. The Luddite movement began in Nottingham and culminated in a region-wide rebellion that lasted from 1811 to 1816.
Decentralised trust: A positive alternative to the AI economic dystopia ?
2
decentralised-trust-a-positive-alternative-to-the-ai-economic-dystopia-13a2d1944f5f
2017-10-04
2017-10-04 01:01:21
https://medium.com/s/story/decentralised-trust-a-positive-alternative-to-the-ai-economic-dystopia-13a2d1944f5f
false
2,416
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Jordane Rollin
Digital Transformation
e813076cfab1
jordane.rollin
2
1
20,181,104
null
null
null
null
null
null
0
null
0
1b6ea11da118
2018-07-24
2018-07-24 14:58:35
2018-07-24
2018-07-24 15:03:37
2
false
en
2018-07-24
2018-07-24 16:09:17
14
13a44cf27ec
5.828616
9
1
0
Many media outlets and consulting firms would like you to believe that Artificial Intelligence (AI) will be automating up to 50% of jobs…
5
The future of commercial AI Many media outlets and consulting firms would like you to believe that Artificial Intelligence (AI) will be automating up to 50% of jobs that exist today within the next few decades. This prevailing notion is due to several factors such as: 1) Budget constraints and replicability; most AI research has been conducted using static datasets without human involvement, resulting in a bias in the industry focusing on human automation, 2) the concept of automation triggers fear and is used as click bait by the media, 3) years of prolific sci-fi culture which shows a variety of dystopian futures where people wage war or are ruled by machines. However, humans will remain an important contributor in our economy for a very long time even if the automation, which has been overstated, truly happens. Humans will not simply do nothing with their lives; the need to communicate with one another, and to create will remain regardless. Indeed, there is a large, mostly untapped, and commercially viable market for AI technology built to have a Human-in-the-Loop component. This means a strong symbiotic and real-time relationship between machine and human to achieve not only productivity gains, but the ability to accomplish tasks that were previously impossible. Throughout human history, people have taken advantage of tools and machinery to increase their productivity, such as farmers using tractors, which resulted in immense economic gains. Now that large sectors of the economy have shifted to the digital world, people are rewarded for their knowledge and ability to transform various types of data into different forms. It can be expected that new tools improving this transformation process are required; adding to a base that includes word processing tools and computer-assisted design software (CAD) for mechanical engineering and architecture. Al techniques have the possibility to do this with even greater accuracy and speed, so the right path towards greater AI commercial viability is designing better and innovative, disruptive software which takes advantage of the best of human and AI strengths. There is clearly the potential of a flourishing market for AI-based User Interface in software. Some of the strengths of humans versus AI have been discussed in the past, however this article is to outline further specific use cases of Human-In-The-Loop system interfaces that can enable them to co-exist and collaborate to produce the most efficient and innovative results. Human-in-the-Loop for Language Generation The auto-correct function in word processors is a good example of AI and human collaboration: through a simple red underline, the human can be alerted that their grammar is incorrect and with a simple right click, the mistake can be fixed. For email writing, Grammarly, or auto-correct on steroids, raised more than $110M which is a testament to investors believing in its commercial viability. The goal of such technology is centered around improving quality of communication. Another interesting use case is SwiftKey Flow which enables typed word recognition by using AI to analyze small, squiggly strokes on a mobile phone virtual keyboard. The result is a text generation that is much faster than standard finger pressing character by character. It does take a bit of practice, but so does learning to type on a full size keyboard with all 10 fingers. The company was acquired by Microsoft for $250M for the commercial value it can provide offering clear performance advantages to the mobile end users. It’s entirely possible that AI could eventually help you fill gaps of knowledge when writing about an unfamiliar topic by tapping on the collective intelligence of other experts on that specific topic. Google SmartReply is an example in this direction although it currently does not really utilize collective intelligence, it mainly simulates how one specific user would respond. A more extreme case of human-AI technology, is a research project called AlterEgo at MIT, where a device on the jaw enables the wearer to internally verbalize a sentence, and through AI the device is able to interpret the micro-movements of the muscles to generate text or pose and receive answers to difficult computational problems. For people who cannot speak, or have other disabilities, it is not difficult to see how this or similar technology could be a successful game changer. This demonstrates a potential new interface for communication between human and machine where AI can serve as the interpreter. Human-in-the-Loop for Reading Eye tracking technology continues to improve and conceivably, an AI layer that recognizes patterns of information attracting your gaze could dynamically modify entire documents on the fly, summarizing and expanding based on active reading. Technology to summarize text is already heavily under development such as the one from Salesforce. Tobii, one of the largest eye tracking technology companies sold approximately $420M USD of products in 2016, clearly demonstrating the profitability and demand for this technology. Curiously, the market has yet to see the combination of text summarization and eye tracking technology but that could be expected to lead to amazing, new opportunities. Lastly, by combining text summarization, eye tracking and language generation, it would even be possible for the AI to identify what you want to read about, show you a small summarization, identify what you actively read from the summary and provide additional details from the original source and enable you to integrate this new knowledge into your own documents with the right context. While it may seem too futuristic, such technology is close to being technologically feasible, and within a decade it could be ubiquitous in assisting people’s ability to communicate. Human-in-the-Loop for Design Design, such as drawing mechanical components in 3D or in architecture, or digital art can take a lot of man-hours. However, a single designer can design intricate mockups using AI-assisted design software, which are currently still in their early stages but very interesting nonetheless. Typically, a user can draw very simple shape like a cartoon, and the AI identify the intent and generate potential high quality versions. Then in turn, the user can select the version that is most aligned with his/her intent, which triggers another set of versions by the AI, and the loop continues until the final acceptable design is reached. Words cannot perfectly describe the process, so we recommend taking a look at these examples. Companies such as Autodesk have very interesting tools in the works, and the future looks bright for AI-assisted CAD designers. While such products are too recent to fully evaluate their commercial viability, tremendous return on investments can be anticipated. Why does AI need humans? For one thing, an AI system is trained by the feeding of specific datasets, but there will always be some missing piece of information which would prevent it from truly attaining a general intelligence with the ability to perform well in many diverse tasks, as opposed to being narrowly trained on a specific task. As such, it is critical for humans to validate the output at all times and guide the AI in the right direction. It is also to prevent unexpected incidents that can turn tragic, such as the case of the self-driving Uber collision. Why do humans need AI? Mainly, the human brain can only process information in relatively small quantities and with a speed of input that matches the senses. It cannot can’t process information outside these ranges, and the AI can serve as an interpreter to convert the sensory signals into a speed and quantity that can be processed by the brain. This way, humans can accomplish tasks that were impossible or even unconceivable before. As assistants for life’s tasks and for the advancement of technology, AI systems can make humans more productive, leading to more free time for meaningful tasks. People could spend more time with their loved ones, stay fit, support their communities, etc. This could positively reshape the relationship to work and hobbies, professional and private time, hopefully for the best. Summary There is a large, ripe market ready for the taking in the area of AI technology that augments humans instead of fully automating them. Even Elon Musk, once champion of full factory automation, now realizes that 100% automation is still not commercially or even technically viable just yet; he recently announced delays regarding the production at Tesla. Through the design of entirely new software built with a Human-In-The-Loop context, humankind will be able to accomplish tasks that are simply unfeasible at the moment, bringing us closer to mastering the natural world in ways we can only imagine at the moment.
The future of commercial AI
101
the-future-of-commercial-ai-13a44cf27ec
2018-07-24
2018-07-24 16:09:17
https://medium.com/s/story/the-future-of-commercial-ai-13a44cf27ec
false
1,443
augmenting collective intelligence
null
eruditescience
null
Erudite AI
info@erudite.ai
eruditeai
ARTIFICIAL INTELLIGENCE,MACHINE LEARNING,SOCIAL IMPACT,EDUCATION TECHNOLOGY
eruditeai
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Patrick Poirier
Co-Founder of Humaniti — Devoted to building artificial individual intelligence technologies to transcend our bio-cognitive limitations.
4f76451031ca
ppoirier.ai
16
5
20,181,104
null
null
null
null
null
null
0
null
0
d0c08339b41f
2018-03-05
2018-03-05 16:50:57
2018-03-05
2018-03-05 16:53:26
1
false
en
2018-03-05
2018-03-05 16:53:26
17
13a49c5f9b3e
1.309434
2
0
0
TWiML Talk 116
5
Scaling Machine Learning at Uber with Mike Del Balso TWiML Talk 116 In this episode, I speak with Mike Del Balso, Product Manager for Machine Learning Platforms at Uber. Subscribe: iTunes / SoundCloud / Google Play / Stitcher / RSS Mike and I sat down last fall at the Georgian Partners Portfolio conference to discuss his presentation “Finding success with machine learning in your company.” In our discussion, Mike shares some great advice for organizations looking to get value out of machine learning. He also details some of the pitfalls companies run into, such as not have proper infrastructure in place for maintenance and monitoring, not managing their expectations, and not putting the right tools in place for data science and development teams. On this last point, we touch on the Michelangelo platform, which Uber uses internally to build, deploy and maintain ML systems at scale, and the open source distributed TensorFlow system they’ve created, Horovod. This was a very insightful interview, so get your notepad ready! Conference Update Be sure to check out some of the great names that will be at the AI Conference in New York, Apr 29–May 2, where you’ll join the leading minds in AI, Peter Norvig, George Church, Olga Russakovsky, Manuela Veloso, and Zoubin Ghahramani. Explore AI’s latest developments, separate what’s hype and what’s really game-changing, and learn how to apply AI in your organization right now. Save 20% on most passes with discount code PCTWIML. Early price ends February 2! About Mike Mike on LinkedIn Mentioned in the Interview Michelangelo Horovod Horovod Github Random Forest Vote on the #MyAI Contest Entries Now!! Register for the Artificial Intelligence Conference here! Check out @ShirinGlander’s Great TWiML Sketches! TWiML Presents: Series page TWiML Events Page TWiML Meetup TWiML Newsletter
Scaling Machine Learning at Uber with Mike Del Balso
2
scaling-machine-learning-at-uber-with-mike-del-balso-13a49c5f9b3e
2018-03-13
2018-03-13 17:09:08
https://medium.com/s/story/scaling-machine-learning-at-uber-with-mike-del-balso-13a49c5f9b3e
false
294
Interesting and important stories from the world of machine learning and artificial intelligence. #machinelearning #deeplearning #artificialintelligence #bots
null
twimlai
null
This Week in Machine Learning & AI
team@twimlai.com
this-week-in-machine-learning-ai
MACHINE LEARNING,ARTIFICIAL INTELLIGENCE,DEEP LEARNING,PODCAST,TECHNOLOGY
twimlai
Machine Learning
machine-learning
Machine Learning
51,320
TWiML & AI
This Week in #MachineLearning & #AI (podcast) brings you the week’s most interesting and important stories from the world of #ML and artificial intelligence.
ca095fd8e66c
twimlai
292
33
20,181,104
null
null
null
null
null
null
0
null
0
30e01424da4e
2018-07-18
2018-07-18 15:49:36
2018-07-18
2018-07-18 15:57:36
3
false
en
2018-07-18
2018-07-18 15:57:36
1
13a4a0b234dc
5.104717
66
2
1
Smart office uses biosensors and machine learning to optimize individual work environments
5
Mediated Atmosphere Smart office uses biosensors and machine learning to optimize individual work environments The atmosphere of a given space — the light, sounds, and sensorial qualities that make it distinct from other spaces — has a marked, quantifiable effect on the experiences of the people who inhabit those spaces. Mood, behavior, creativity, sleep, and health are all directly impacted by one’s immediate surroundings. In the workplace, atmosphere can influence productivity and relationships, as well as overall employee satisfaction and retention. Recent studies have identified a decline in workplace satisfaction — particularly in the knowledge economy, where distraction and disengagement can cost billions of dollars in lost productivity and employee turnover. Mediated Atmosphere, a project by the Responsive Environments group at the MIT Media Lab, seeks to improve both wellbeing and productivity in the workplace by improving the workplace atmosphere at an individual level. Using modular, real-time control infrastructure with biosignal sensors, controllable lighting, projection, and sound, Mediated Atmosphere creates immersive environments designed to help users focus, de-stress, and work comfortably. A “smart office” with biosensors and machine learning With the boom of Internet of Things technologies over the last few years, then-master’s student Nan Zhao noticed that the many lighting solutions, wireless speakers, and home automation platforms on the market lacked a multimodal quality: synchronizing light, sound, images, fragrances, and thermal control in a meaningful way. Also missing in most available “smart” home and office products is a basis in physiology — platforms that incorporate research on the impact of atmospheric scenes on cognition and behavior. For this project, Zhao drew on existing research showing the positive effects of natural views and sounds on mental state, as well as the effects of light and sound on mood, alertness, and memory. In the course of this research, however, Zhao kept coming to the same conclusion: “It’s not one size fits all. People need a place that is fascinating, that gives them a feeling of being away, and is rich but predictable. However, this place is different for different people. With our approach, we want to create a personalized experience.” Comprising a frameless screen (designed with a special aspect ratio so it doesn’t feel like watching TV), custom lighting network, speaker array, video projection, and both wearable and contact-free biosignal sensors, Mediated Atmosphere synchronizes and controls numerous modalities. Zhao and her collaborators developed a new approach for controlling the system: a control map that compresses a complex set of input parameters to a simplified map-like representation. The “compass points” of this map are abstract control dimensions, such as focus or restoration. This way, rather than worrying about light levels or sound sources, users can simply tell the system what they want based on how focused or relaxed they want to be. The biosignal sensor stream computes a focus and restoration indicator based on measures developed and evaluated by Zhao and her team. Using these indicators, Mediated Atmosphere can label what specific atmospheric scenes mean for the user, and learn how to automatically trigger changes based on a user’s actual responses and activities. Customized workspace The smart office concept is designed to self-regulate on the basis of the user’s activities and physiology. Using biosignal sensors to track heart-rate variability and facial expressions, the prototype both responds to the user’s moods in real time and tracks responses. A user study published in Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies in June 2018 found that the Mediated Atmosphere smart office prototype had a positive effect on occupants’ perceptions and physiological responses. “We imagine a workspace that, when asked, can instantly trade the engaging focus of a library with the liberating sensation of a stroll through the forest,” explains Zhao, the first author on the paper. “We want to create an ‘environment player’ that can recommend or automate your space similar to how Spotify or Pandora gives you access to a world of music. We want to help people to manage their day by giving them the right place at the right time.” The study of 29 users offered five different ambient scenes, ranging from forest streams to bustling coffee shops, measuring how the environment influenced participants’ ability to focus and restore from stress. A second study with nine subjects and 33 scenes, published in Zhao’s thesis, looked at how well the user interface worked in applications where the choice of environments was driven by sensors. In future iterations, Zhao hopes to give users the ability to record their own personal favorite places and upload them into the system, in addition to the built-in options. Mediated Atmosphere. Credit: Nan Zhao Zhao is working with a number of industry experts to hone both the technology and the experiential effectiveness of Mediated Atmosphere. Media Lab alumna Susanne Seitinger, a lighting expert at Philips, worked with Zhao on the lighting installation. Steelcase has advised Zhao on designing for workplaces. International Flavors and Fragrances, a Media Lab member company, is supporting the team’s efforts to add an olfactory display into the latest prototype. Most recently, member company Bose has been supporting the work and helping to take the prototype to the next level; the next iteration will be a modular system that can be installed in any existing workspace so Zhao’s team can conduct experiments on this technology in the wild. Lee Zamir, director of the BOSEbuild team, is enthusiastic about Mediated Atmosphere’s potential to help redefine the workspace. “The Mediated Atmosphere project has the potential to improve and rethink the work environment,” he says. “We go to work not just to make a living, but to be challenged, to accomplish, to focus, and to connect with others to achieve great things. When we are able to do this, when we have a ‘good day at work,’ it improves all the other parts of our lives. We carry that sense of purpose and progress from our workday with us.” In addition to the next phase of research in office environments, Zhao is also creating a smaller, modular system that could be installed in any office or even in a home office. The team is exploring more sensory modality such as thermal control, air flow, and scent. Atmospheric Scenes. Credit: Nan Zhao Future office Zhao envisions a future office where employees’ workstations come equipped with Mediated Atmosphere platforms, but the concept is a long way from being ready to market or scale. One major challenge is to measure impact reliably during real work scenarios without burdening the user; to that end, Zhao is developing a contact-free sensor system to remove the wearable component. Another difficulty is creating customizable installations that fit into different sizes and types of office spaces, allowing colleagues to each have their own Mediated Atmosphere workstation without disrupting one another. The team is collecting data and doing image-based analysis using machine learning tools to address this challenge. But perhaps the challenge Zhao takes most seriously is that of adding real value. “The same technology that can create a memorable, wonderful, stimulating experience can also create an irritating, elevator-music type of experience,” she says. “It takes artistic intuition and empathy to create the former. That is also why personalization is so important.” This story was originally published on MIT News.
Mediated Atmosphere
345
mediated-atmosphere-13a4a0b234dc
2018-07-18
2018-07-18 17:12:00
https://medium.com/s/story/mediated-atmosphere-13a4a0b234dc
false
1,207
News, ideas, and goings-on from the Media Lab community
null
mitmedialab
null
MIT MEDIA LAB
info@media.mit.edu
mit-media-lab
EDUCATION,STEM,DESIGN
medialab
Research
research
Research
17,602
MIT Media Lab
Ideas and goings-on from the Media Lab community. Find our blog archive at http://blog.media.mit.edu/
693176a2e08c
medialab
289,711
460
20,181,104
null
null
null
null
null
null
0
null
0
f3265ac69af4
2018-09-14
2018-09-14 14:48:45
2018-09-14
2018-09-14 14:54:21
2
false
en
2018-09-14
2018-09-14 14:54:21
9
13a4cf5f579d
2.372013
0
0
0
For Immediate Release: The Chatbot Conference, the premier conference dedicated to the coming generation of chatbots, AI, voice and virtual…
5
Seed Co-Founders to Keynote The Chatbot Conference on September 19th For Immediate Release: The Chatbot Conference, the premier conference dedicated to the coming generation of chatbots, AI, voice and virtual assistants, today announced that Nathan Shedroff and Mark Stephen Meadows, co-founders of Seed Vault, an independent, open-sourced platform dedicated to the design and development of conversational user interfaces (CUIs), will lead its keynote on September 19, 2018 in San Francisco. Focusing on the democratization of AI, the development of trusted systems, and how blockchain can be used to promote accountability, Shedroff and Meadows’ talk describe an antidote to the monopolization of our personal data by corporations like Facebook, Amazon, and Google. As Mark Stephen Meadows, Co-founder of SEED said, “AI and Conversational User Interfaces are too critical to the advancement of digital services and economies to leave to a few global corporations bent on owning all data and controlling all relationships. We need to be able to trust our interactions with bots, AI, and the organizations and people they represent.” Nathan Shedroff, SEED CEO, added, “SEED is leveling the playing field with an open source, independent alternative to the corporate monopolies that currently dominate CUI and AI technologies. SEED uses blockchain to promote trust, accountability, and a thriving economy for organizations, bots, and people alike, giving back control of their data, as well as putting in place a monetization model that enables people to be paid for their contributions to AI systems.” About Nathan Shedroff Nathan Shedroff is CEO of Seed Token, an independent platform for the development of Conversational User Interfaces (CUIs). By enabling people to speak to AI systems, CUIs represent a paradigm shift in the way we interact with computers, digital services, and each other. Previously Nathan founded and led the Design Strategy MBA program at California College of Art. One of the pioneers of Experience Design, Interaction Design, and Information Design, Nathan is the author of Blind Spot, Experience Design 1, Design is the Problem, Making Meaning, and Understanding Computers, among others. About Mark Stephen Meadows Seed Co-Founder Mark Stephen Meadows holds six patents including one for systems and methods for cryptographically secure transactions using voice and natural language processing systems. With 18 years in AI, 22 in VR, and 7 years in blockchain he has designed and developed AI applications at some of the world’s top research labs including Xerox-PARC and SRI. Since 2011 Mark has led Botanic.io, which is pioneering the development of AI, blockchain, and conversational interfaces in finance. Botanic customers include Logitech, BMW, Sense.ly, and start-ups in the US, Australia, and China. ABOUT SEED SEED is an open, independent, and decentralized marketplace for developers, publishers and users of conversational user interfaces (CUIs) or “bots”, that democratizes AI. The SEED platform provides development tools, intellectual property, and a tokenized network for delivering front-ends to AI technologies. To learn more you can find SEED on Telegram, or visit the SEED website and sign up for email updates at https://seedtoken.io/ If you have specific questions about the project, please contact ask@seedtoken.io. You can also find SEED on Discord, on Twitter, on Redditand on Medium.
Seed Co-Founders to Keynote The Chatbot Conference on September 19th
0
seed-co-founders-to-keynote-the-chatbot-conference-on-september-19th-13a4cf5f579d
2018-09-14
2018-09-14 14:54:21
https://medium.com/s/story/seed-co-founders-to-keynote-the-chatbot-conference-on-september-19th-13a4cf5f579d
false
527
SEED is an open and independent marketplace for developers and deployers of conversational user interfaces (CUIs) that democratizes AI.
null
null
null
seedtoken
steve@seedtoken.io
seedtoken
BOTS,CHATBOTS,AI,CONVERSATIONAL UI,CONVERSATIONAL INTERFACES
seed_token
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
SEED Token
Designed explicitly to enable an open, global, trusted bot economy.
b4e7d32f2915
seedtoken
66
24
20,181,104
null
null
null
null
null
null
0
null
0
ec10e05abbed
2018-05-10
2018-05-10 03:15:10
2018-05-10
2018-05-10 03:21:45
4
false
zh
2018-05-10
2018-05-10 03:28:11
9
13a618290a72
2.562
1
0
0
Cortex自上线以来,获得了Ai领域、区块链领域从业者、爱好者、投资者的广泛关注,中英文社群也在快速成长中。
5
Cortex新一轮空投正式启动 价值百万CTXC等你来拿! Cortex自上线以来,获得了Ai领域、区块链领域从业者、爱好者、投资者的广泛关注,中英文社群也在快速成长中。 为了进一步促进社区良性成长,Cortex项目官方发布新一轮大规模空投活动。 规则如下: 中文社区投稿奖励 Cortex项目团队联合链向财经,推出专栏征集Cortex主题相关文章的活动。欢迎大家到链向财经发布Cortex项目动态、对Cortex项目的看法、token的行情分析、Cortex项目的技术分析或未来发展等内容,链向财经内容团队会进行审核,一旦审核通过会发布到Cortex项目库专栏下。投稿者将会获得100人民币对应的等值CTXC(按照发币当天市价兑换)。 活动链接、PC端:http://t.cn/R3wsaJn 移动端 Twitter、Telegram群组 reddit的社区奖励 2018年5月9日,Cortex正式上线了Twitter、Telegram群组、reddit相关的社区bounty奖励系统。 规则如下: 每天奖金池包括750 CTXC。针对不同的任务(如关注/点赞/转发/回复Twitter、telegram邀请好友/积极发言、reddit订阅/参加投票),会有不同的积分奖励。当日完成任务获得的积分,按照任务积分占当天总积分的比例瓜分奖金池的CTXC。 具体规则可参考空投页面详情介绍。 活动链接、PC端:http://t.cn/R3AvZ9V 移动端 真诚的欢迎大家通过以上任何一种方式参与到Cortex社区的建设中来,让Cortex社区健康、有序、良性的快速成长。 联系我们 网站:http://www.cortexlabs.ai/ Twitter:https://twitter.com/CTXCBlockchain Facebook:https://www.facebook.com/cortexlabs/ Reddit:http://www.reddit.com/r/Cortex_Official/ Medium:http://medium.com/cortexlabs/ 电报:https://t.me/CortexBlockchain 中国电报:https://t.me/CortexLabsZh
Cortex新一轮空投正式启动 价值百万CTXC等你来拿!
23
cortex新一轮空投正式启动-价值百万ctxc等你来拿-13a618290a72
2018-05-10
2018-05-10 19:53:42
https://medium.com/s/story/cortex新一轮空投正式启动-价值百万ctxc等你来拿-13a618290a72
false
25
AI on Blockchain - The Decentralized AI Autonomous System
null
CTXCBlockchain
null
Cortex Labs
support@cortexlabs.ai
cortexlabs
AI,BLOCKCHAIN,CRYPTOCURRENCY,CTXC,CORTEXLABS
CTXCBlockchain
区块链
区块链
区块链
617
Ëazon
null
254da0c8ee50
Eazon
3
1
20,181,104
null
null
null
null
null
null
0
null
0
null
2018-06-20
2018-06-20 22:06:28
2018-06-20
2018-06-20 23:02:32
6
false
zh-Hant
2018-06-20
2018-06-20 23:02:32
6
13a6be41b4be
1.274528
0
0
0
今日主題:變分自編碼器
2
Day 31 — Variational Autoencoder 今日主題:變分自編碼器 參考資料 NEURAL NETWORK ZOO PREQUEL: CELLS AND LAYERS Towards Data Science Variational Coin Toss 原始論文:Auto-Encoding Variational Bayes Variational Autoencoders Explained 與作者的Github 筆記 是一種Autoencoder,與基本的AE的結構比較圖 [1] 圖例 跟AE一樣,VAE會壓縮輸入資料產生編碼。但壓縮的是資料的機率而不是特徵。 輸入會是input sample的近似機率分布而不是直接給資料本身 最根本的想法節錄自[1] The basics come down to this: take influence into account. If one thing happens in one place and something else happens somewhere else, they are not necessarily related. If they are not related, then the error propagation should consider that. 結構圖也可以用[5]的這張圖表示 至少貓咪很可愛。 在對比兩種結構圖之後,好像比較容易理解它在幹嘛了。中間的Encoder/Decoder Network應該仍然用的是感知機層,只是在中間多加上了計算機率分布的mean/sd的計算層。 原始論文提供了VAE mini-batch版本的演算法 損失函數應該一樣是定義在輸入跟輸出結果的差異上。訓練的方式應該也是跟標準AE一樣採用倒傳遞法訓練Encoder+Decoder,訓練完畢之後再將Decoder拿掉就是最後可以拿出來使用的成品。 [5]的最後有利用VAE針對MNIST手寫數字資料集做測試,結果如下 作者說這個成品只用了他自己的普通筆電15分鐘就做完了,看來是個蠻實際可行的模型。 作者已經把他寫的程式碼分享在Github上。
Day 31 — Variational Autoencoder
0
day-31-variational-autoencoder-13a6be41b4be
2018-06-20
2018-06-20 23:02:33
https://medium.com/s/story/day-31-variational-autoencoder-13a6be41b4be
false
86
null
null
null
null
null
null
null
null
null
Machine Learning
machine-learning
Machine Learning
51,320
Falconives
null
250d8013fad2
falconives
11
19
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-10-11
2017-10-11 12:45:40
2017-10-12
2017-10-12 22:09:42
4
false
en
2017-10-13
2017-10-13 07:21:58
1
13a8582e5ae2
4.111321
2
0
0
When I retrospect about interesting Sci-Fi movies of last two decades, a scene from Minority Report (2002) always pops up where Tom…
5
Future of Marketing with Augmented Reality When I retrospect about interesting Sci-Fi movies of last two decades, a scene from Minority Report (2002) always pops up where Tom Cruise-protagonist visits a shady place which is a sort of black market and apparently all the shops are providing services which are based on either augmented or virtual reality. Although it could be a figment of imagination for some ambitious graphic designers back then but it can’t be less than any reasonable real time touch-points for new age marketers. Augmented reality is one of the latest technological means for companies and brands to deliver their persuasive techniques to tech-savvy audience. And, as we speak today we are not looking at alien or complex format of technology for the target audience with a decent smartphone in hand. We all are initiated when it comes to AR either through filters like Facebook, Snapchat and other social media platforms or games like Pokemon Go which had been a recent hype especially when one observes the eventual stock-price hike for Nintendo. Last Five years commercial application of such technologies have helped in building the trend of developing utilities which are providing screen-enhanced information. Augmented reality is sort of overlay of such personalized, accessible and appealing set of information on to the real world which allows users to see virtual elements or information while looking at real time entities. In current scenario in India, most basic commercial application of AR so far has been for events and activation programs. Brands like Coca Cola, Tata Tiago, Renault KWID in India are using AR based personalized experiences in on-ground mall spaces and public event spaces for mass. Mahindra back in 2012 used AR in auto Expo, while launching XUV 500, to let people over there experience a virtual cheetah which was a visual treat and gained an impressive traction. So, Gaming and entertainment is all what AR can do to get some eyeballs to brands? But, that seems the current or feasible utility-end of it. Some sorts of opportunist take at cashing in “shiny object syndrome” to uninitiated? At the same time what appears to me as a challenge and opportunity as well that how it can be utilized as a tool for experiential marketing effectively. Brands can look at AR as an extra-hand in perceiving a better brand recall value, positive word-of-mouth, and cultivating a repeat buying behavior by attractive and easy purchase journey. If we look at examples, real estate (CommonFloor) and consumer care have started vouching their new touch-points to consumers. We have Lenskart providing an AR service to try out different samples of frames for consumers on App. While it adds a touch of novelty and also helps brand with a unique brag-worthy proposition thereby churning a decent word-of-mouth, brands can start exploring new options where AR can be a bridge between digital market place and a traditional brick and mortar shop. A mother with six-year old enters in a toy shop of let say Hasbrow and hands over the phone with the app of the brand to the child. Let’s say whenever kid rows the app over toys available in the shop, they (products let’s say a teddy or Captain America) start talking in personified manner through app or demo of the toy gets projected on phone screen. Such example of engagement will make retain customers and also provide tangible shareable property amongst prospective target group thereby giving another means to socialization. Apart from such in-store better consumer experience (Walmart tried it long time back), another prospective marketing landscape with AR would be in ideating and introducing interactive advertisements. And, Indian tech- startups are joining the AR bandwagon with creative solutions to offer. Startups like Shopsense are helping apparel stores where customers can zero down their choice by virtually trying out on screen consisting in-store inventory. The AR market internationally is projected to reach $117.4 billion (at compounded growth rate of 76%) by 2022 according to MarketsandMarkets report titled “Augmented Reality and Virtual Reality Market — Global Forecast to 2022”. “The increasing demand for AR and VR technology-based products in various verticals such as consumer, aerospace, defence, commercial and medical is responsible for the market growth,” said the MarketsandMarkets’ report. The Indian AR & VR market is projected to grow at a CAGR of over 55% during 2016–2021. Increasing AR usage for enhancing gaming experience, growing penetration of head-up displays in the automotive sector, rising use of AR in pico projectors, etc., are driving the AR & VR market in the country. AR opens up unexplored field for marketeers as it provides new space for three-dimensional thinking and idea execution. It is turning out to be an immersive platform of storytelling for brands with more objects to interact for consumers and edge-cutting real time experiences. Brands like Lego and IKEA are allowing customers to scan the catalogue and decide where the products would fit in their living spaces or not. Surprisingly sitcoms like Black Mirror are doing a bit much than real time product developers and marketing professionals (mainly because of artistic liberty and practicality of such products and services in reel life) but there’re potential touch-points for companies to communicate their brand-attitudes and opportunities for agencies of showcasing their reinventing capabilities for brands. Pritesh Desai is a final year student at MICA pursuing PGDM-C.
Future of Marketing with Augmented Reality
4
future-of-marketing-with-technologies-like-augmented-reality-13a8582e5ae2
2018-05-09
2018-05-09 09:43:47
https://medium.com/s/story/future-of-marketing-with-technologies-like-augmented-reality-13a8582e5ae2
false
904
null
null
null
null
null
null
null
null
null
Augmented Reality
augmented-reality
Augmented Reality
13,305
The Marketing Orbit
This blog is a platform for MICAns to express their opinions on everything Marketing! Curated by Team MiCompass which is the Marketing Club of MICA, Ahmedabad.
91214097a7f9
marketingorbit
70
15
20,181,104
null
null
null
null
null
null
0
null
0
347f85c68611
2018-02-07
2018-02-07 21:43:43
2018-02-07
2018-02-07 21:50:48
3
false
en
2018-02-07
2018-02-07 21:50:48
1
13a9f23b1bf8
5.225472
2
0
0
Hello World! The bots are here and they are changing the way in which humans experience life. The study of usage of mobile devices globally…
5
Driving Businesses through Customer Engagement Hello World! The bots are here and they are changing the way in which humans experience life. The study of usage of mobile devices globally reveals some breakthrough statistics. People have quite clearly taken advantage of the ever growing social media but they have also specified their preferred mode of interaction. Let’s go to some visual cues straight away! These stats from Business Insider show the exponential rise of Messaging & Social vs other apps. The average digital time spent in ‘Social and Messaging apps’ grew by a stark 394%, in contrast to the average year-on-year growth of 69% for mobile app usage, from 2015 to 2016. No wonder investors are going bullish over this trend! Customers are widening their social presence and experiences and generating crucial data on the fly. Any business that wants to stay competitive has to be directly engaged to each and every customer in this data-driven age, whatever the platform. What better than penetrating spaces where your customers slay most of their time! The rise of connected devices and Internet of Things (IoT) has further given vent to the fire. IoT has had a ripple effect on the world. Its fundamental goal is to connect most of our possessions as a cohesive and interactive net. Wearables, Connected security systems, kitchens, connected cars are among a long scroll of items that can interact amongst themselves, to do most of our routine things. Amazon Echo and Fitbit One are two examples of public use. While in the industrial segment, DHL’s IoT Tracking and Monitoring and Cisco’s Connected Factory are prime examples. The promise and surge in usage is enthralling at this point! There are around 28.4 billion connected devices as of now and is estimated to hit 50 billion by 2020. To give you a comparison scale, there are approximately 7.5 billion humans on Earth. Systems can now interact with each other and businesses can reach customers through mobile devices. Yet the way a customer communicates his/her interest, grievance, etc. is far from the most preferred channel — messaging apps and services. Enter Chatbots! Rather than opening an app or a browser and searching for a place to stay, while on a one-day mission to some far-off city, one would just want to make are quest for a room as a simple conversation. “Are there any rooms nearby with good ratings along with some good lounges around?” And, bam! The answer’s there along with a nice little chat. Using one’s own words to search, transact and interact brings a whole new dimension in consumer experience. Chatbots are programs which twist the existing click and tap based User Interface to that of conversations. You would have already heard about or used them– Facebook’s M, Amazon’s Alexa and Apple’s Siri are all versions of chatbots. Chatbots infer from the context, and enable users to access various services like booking flights, etc. while they are still on messaging/calling apps like Facebook messenger or Skype. Pure chat platforms like WhatsApp will soon join the bandwagon. Chatbots are set to disrupt businesses because, along with numerous benefits, they give consumers the most natural of interfaces with AI i.e. language. The kind and speed of interaction and response chatbots have, will undoubtedly shape how your business interacts with customers, and thus drive your growth. That’s where Saarthi comes in! Empowering businesses and consumers Saarthi is a multi-lingual Conversational Agent from Gamut Analytics, an AI & Deep Learning driven business consulting company founded in 2016. It creates products based on algorithms to solve problems from a myriad of backgrounds such as speech recognition, customer relation management, fraud detection systems, pricing and recommendation engines, campaign optimizers and so on. Like each of its products, Gamut Analytics has crafted yet another data masterpiece. Saarthi automates service and support to customers and enables handling of complex queries, transactions and decisions without sacrificing speed and accuracy. It transforms the user interface and engages with each customer by talking in his/her own personal flavour. Saarthi_Demo_Bot helping close a potential car loan deal. Please note that this is just intended for demonstration purposes and has no association with any entity or individual. Besides that; ¨ It is channel-agnostic, allowing flexibility and making products available to customers quickly and with ease. ¨ The bot can easily integrate with Facebook Messenger, Skype, Kik, etc. ¨ It is available 24*7 and has no bad days or mood swings ¨ It can communicate simultaneously to any number of clients, thus scaling according to demands of the day. ¨ It uses Machine Learning and Natural Language Processing to capture queries and respond in a number of languages (For now, but not limited to, Hindi, English and Hinglish — both speech and text). Saarthi’s engine draws upon huge domain-relevant databases and is embedded with FAQs to make incident management, grievance redressal and general queries available fast and anytime a customer wants .It also uses the data to effectively understand the customer and the way he/she wishes to interact with the business. Still better, it learns with time thus increasing the ‘personal element’ gradually. Saarthi will be drastically changing the face of B2C and B2B operations. Companies, that have the ability to identify and create personas unique to each consumer, can anticipate their needs and improve customer service. The more you understand each customer, the better your customer loyalty will be. Saarthi enables a customer-centric business model where business decisions like investments, new products, expansions, etc. are driven by insight and not just intuition. Data gathered from conversation threads can be used for ¨ increased targeted marketing, ¨ accurate brand representation, ¨ cross-sell/up-sell opportunities, ¨ reduction in process and expenditure waste ¨ and discovering new leads by tapping into channels across geographies. Businesses will profit from operational cost reduction and exponentially increase, revenues and the all-crucial bottom line. Moreover, with the constant stream of data generated through Saarthi’s conversation threads, you can track the vital KPIs with the help of trackers or dashboards and have a live health-monitoring of your business and adapt to the ever-changing whims and needs of the market in one exploratory and insightful glance. Gamut Analytics also intends to enter into strategic analytics/consulting partnerships to provide constant augmentation to business by leveraging bot data. Saarthi is set to span a diverse range of domains from banking to hospitality to medical practices. Pretty soon, we will be reserving dinner tables, booking flights, fixing appointments or ordering a late-night craving through Saarthi. We will be talking to Saarthi because of the experience, speed and convenience. So, businesses need to up their game and cash in on this ‘winner’. No half measures- All or Nothing! And about the bullishness of investors, have the bots just recently become relevant to their ventures? No. The boom sparked back in 2015. The next year, global funding for startups built on AI, saw some 60% more. The world of AI is thrilling for venture capitalists as it presents a next-age platform for both setting up and growing businesses whilst simultaneously providing convenience to customers, which is essential for sustainability. Businesses now adapt to technology at a dizzying pace to keep up with the transformations in customer behavior and have started integrating with messaging services to reach them directly. Facebook is certainly pushing the creation of chatbots built on top of Messenger and many companies are leveraging it. CNN, Wall Street Journal, HP and Health Tap are examples of some. At this turn of time and trend, enterprises need to go out all guns blazing or live to be haunted by regret.
Driving Businesses through Customer Engagement
2
driving-businesses-through-customer-engagement-13a9f23b1bf8
2018-04-11
2018-04-11 01:46:59
https://medium.com/s/story/driving-businesses-through-customer-engagement-13a9f23b1bf8
false
1,239
Saarthi is a virtual assistant to run all your errands, optimally! We also offer Enterprise Development services to develop bots across popular communication channels.
null
thesaarthi
null
Saarthi.ai
hello@saarthi.ai
saarthi-ai
CONVERSATIONAL INTERFACES,ARTIFICIAL INTELLIGENCE,INTERNET,BOTS,AMBIENT INTELLIGENCE
saarthi_ai
Chatbots
chatbots
Chatbots
15,820
Sangram Sabat
Chief Growth Officer — Saarthi.ai{ Unlocking Internet for 536 Mn 🇮🇳language 🗣️ } | 📈Hacker | AI 👁️‍🗨️| Past- Sr.Consultant (Citi)
48f1d6910ca5
sangramsabat
56
61
20,181,104
null
null
null
null
null
null
0
null
0
null
2017-10-05
2017-10-05 03:57:49
2017-10-17
2017-10-17 15:43:03
9
false
ko
2018-01-03
2018-01-03 06:00:08
1
13aa46bf8387
9.358
1
0
0
Written by Ray Kurzweil
3
[Book Review] 마음의 탄생 Written by Ray Kurzweil 마음의 탄생. 크고 두껍다. 0. Table of Contents Why this book? About the author Glossary Key sentence Summary Evaluation Commitment 1. Why this book? 애청하고 있는 지대넓얕의 143회 ‘AI와 영생’편. 채사장님이 ‘레이 커즈와일’의 책 <특이점이 온다.>를 소개했다. 10년이 된 책임에도 불구하고, 저자가 그리는 미래는 상당히 과감하면서 매혹적이었다. ‘특이점-Singularity’는 컴퓨터의 지능이 인류의 지능을 앞지르는 시점을 말하는데, 이 책의 제목은 그 날이 멀지 않았음을 의미한다. <마음의 탄생>은 저자의 특이점에 대한 생각들의 연장선 상에서, 컴퓨터로 인간의 마음을 구현할 수 있을지를 다룬 가장 최근 저서이다. (2012년) 2. About the author 그의 책은 SF소설 같은 느낌이 든다. 그래서 어쩔 수 없이 서두에 레이 커즈와일의 이력을 나열한다. (안 그러면 너무 약 파는 느낌) 21세기 버전의 진시황 느낌. 영생을 추구하신다. 1948년: 많은 천재가 그렇듯이 뉴욕에서 유대인으로 태어난다. 15세-17세: 클래식 음악작품을 분석한 뒤 비슷한 스타일로 합성하는 패턴인식 소프트웨어 프로그램을 만든다. CBS 방송에 출현해 이 프로그램이 작곡한 악보를 피아노로 연주한다. 발명상을 휩쓸고, 백악관에 초청 받아 린든 존슨에게 격려 받는다. 20세: MIT에 개설된 프로그래밍 수업을 1년 반 동안 모두 클리어하고, 수천 개 대학의 모집요강을 DB화하여 입시지원자 개개인에게 맞는 대학을 제안하는 프로그램을 제작한다. 매년 로열티를 받는 조건으로 강남 아파트 값(현 시세 7.5억원)을 받고 매각한다. 26세: 평판스캐너기술과 음성기술을 결합하여, 맹인을 위한 Kurzweil Reading Machine을 출시한다. 전미 스케일의 유명인이 된다. 4년 후 제록스에 매각하고, 제록스의 기술자문이 된다. 36세: 스티비 원더에게 영감을 얻어 신디사이저를 만든다. 음질테스트에서 그랜드피아노에서 나오는 소리와 구분할 수 없었다고 한다. 바로 세계적인 악기 브랜드가 된다. 영창악기에 매각한다. — 미래학자, 발명가, 작가로서 다양한 활동 — — 64세: 구글의 ‘머신러닝과 언어처리 프로젝트’ 책임자가 된다. 아이폰 시리 개발의 핵심 역할을 한다. 3. Glossary 3–1) 신피질(Neocortex): 감각 지각, 시각적 대상 인식, 추상적 개념 인식, 동작 제어, 공간 지각, 합리적 판단, 추론, 언어 구사에 이르기까지 광범위한 부분을 관장하는 뇌의 겉표면(outer surface). 4. Key sentence “뇌(신피질)와 똑같은 방식으로 작동하는 인공지능을 만들지 못할 이유가 전혀 없다.” 5. Summary 인간이 생각이라고 간주하는 모든 활동은 뇌의 신피질에서 관장한다. 따라서 신피질의 작동 원리를 알게 되면, 베일에 싸인 인간의 생각의 메커니즘을 알 수 있을 뿐만 아니라 디지털 뇌도 구현이 가능할 것이다. 5–1) 신피질의 작동 원리: 패턴인식 마음이론 저자는 아래의 생각실험을 통해 다음과 같이 후려친다. 신피질 작동원리: 패턴을 인식하여 정보를 처리한다. (=패턴인식 마음이론) 우리 기억은 순차적이며 그 순서는 정해져 있다. 입력된 순서대로만 출력할 수 있다. 우리는 기억의 순서를 거꾸로 뒤집지 못한다. (e.g. 알파벳이나 주민번호를 거꾸로 외우기 어렵다.) 뇌에는 이미지, 비디오, 소리를 기록하고 저장하는 장치가 없다. 우리 기억은 패턴의 나열로 저장된다. (e.g. 방금 지나간 여자의 모습을 묘사하기 어렵지만, 여러 사진 중에 지목하는 것은 쉽다.) 우리 뇌는 패턴을 인지한다. 정보의 일부분만 인지하더라도, 인지 능력은 패턴의 변하지 않는 특징을 명확하게 감지한다. => 저자는 인간만의 독보적인 창의력이나 예술적인 능력도 제한된 시간에 패턴인식기를 얼마나 많이 작동시키는 지에 달렸다고 본다. 우리 뇌는 정보가 일부 변형되더라도 패턴의 특징으로 감지할 수 있다. 패턴인식을 통해 하위 레벨 패턴(A)부터 상위 레벨 패턴(Apple 또는 문장)으로 계층적 사고를 한다고 가정한다. 5–2) 신피질의 하드웨어 스펙 직접 계산해보았다. 신피질의 연산 속도: 100개의 뉴런으로 구성된 패턴 인식기는 1초에 3*10⁵번 연산(Transaction)한다. 패턴인식기 600개가 모여 피질기둥을 이루고, 500,000개의 피질기둥이 모여 신피질을 이룬다. 신피질에서 패턴인식기들은 병렬로 연결되어 있으므로, 초당 9*10¹³번 연산이 가능하다. 신피질 메모리: 72바이트 용량을 가진 패턴인식기가 3억 개 있으니, 20 기가 바이트 정도로 추정된다고 한다. 블루레이 고화질 영화 한 편 수준의 메모리라니.. 이 부분은 논란의 여지가 있겠다. (오래된 기억의 경우, 기억이 엄청 흐릿하다가 겨우 생각나는 경우가 있다. 안 쓰는 기억들이 신피질보다 먼 장기기억세포에 저장되다가 로딩되는 느낌이다. 아이폰으로 치면 아이클라우드 느낌.) 신피질은 정보 전달 속도에서의 생물학적 한계를 갖지만, 엄청난 병렬구조(3¹⁰ 개의 뉴런)와 메모리와 CPU가 붙어있는 구조로 아직까지 컴퓨터에 우위를 유지하고 있다. 5–3) 디지털 뇌를 만들 수 있는가? 컴퓨터 연산 속도: 뇌를 기능적으로 시뮬레이션하려면 초당 10¹⁴~10¹⁶번 연산을 할 수 있어야 한다. 일반 컴퓨터는 아직 신피질에 많이 못 미치는 수준이지만, 일본의 K컴퓨터는 이미 초당 10¹⁶번 연산을 해낸다고 한다. 인텔은 최근 circuit board 공간 문제 해결을 위해 3D 프린팅을 활용해서 쌓는 방식으로 제조를 하고 있다고 한다. 일반 컴퓨터로도 신피질의 연산 속도를 능가하는 시점이 얼마 남지 않은 것으로 보인다. 메모리: 인간의 메모리가 20기가 바이트가 훨씬 넘는다고 해도 컴퓨터를 이길 수 없을 듯.. 이 부분은 컴퓨터가 넘사벽이다. 일본 K 컴퓨터. 신피질에 대적하려면, 이 정도는 되어야 하나보다.. 5–4) 수확가속법칙에 따라 정보기술 발전은 급가속! 수확가속법칙은 정보기술이 예측가능하고 기하급수적인 궤도를 따라 발전한다는 것이다. 아래 그래프를 보면, $1000불짜리 컴퓨터의 연산능력 상승곡선을 로그함수로 나타냈다. 2010년 경에는 쥐 한 마리 뇌의 연산능력에 불과하지만, 2030년 경에 사람 한 명의 연산능력을, 2050년 경에는 전 인류의 연산능력을 능가할 것으로 예측하고 있다. 로그함수 그래프 내에서도 y=x² 기울기.. 과연? <특이점이 온다.> 내용을 역자가 친절히 요약해주셨다. 5–5) 그래도 복제할 수 없는 인간만의 의식이 있지 않을까? 저자는 그런 거 없다고 한다. 의식이나 마음도 뇌의 작용일 뿐이라고 하며, 최근 연구 사례들을 소개한다. 5–4–1) 분할뇌 실험 좌뇌와 우뇌를 연결하는 뇌들보가 절제된 사람이 있다. 이 사람의 좌뇌가 명령을 해서 오른손을 움직였더니, 후에 우뇌가 자기가 명령을 해서 오른손이 움직인 것으로 합리화(작화증-이야기를 꾸며냄) 했음이 밝혀졌다고 한다. 5–4–2) 리벳 실험 두피에 EEG(뇌전도) 전극을 붙인 실험 참가자들에게 ‘움직이고 싶은 욕구나 충동을 처음 인지하는 순간’에 버튼을 누르게 했다. 참가자들이 버튼을 누른 시간은 실제 움직인 시간보다 0.2초 정도 앞섰다. 하지만 EEG신호를 살핀 결과, 운동피질이 행동을 촉발시킨 시간이 욕구를 보고한 시간보다 0.3초 앞섰다. 결국 행동을 하기로 결정 내리기 전에, 운동피질이 그 임무를 수행하기 위해 준비했다는 뜻이다. 6. Evaluation 6–1) 사고실험에 기반한 예측은 아직.. 저자의 패턴인식 마음이론은 사고실험에 기반하기 때문에, 아직 나에게 확실한 설득력을 갖지 못한다. 아인슈타인이 사고실험으로 상대성이론을 도출했지만, 그 성과는 후대 과학자들이 실험적으로 검증하면서 더욱 빛을 받게 되었다. 저자도 이 책의 예측이 검증이 되고, 패턴인식기를 활용한 컴퓨터가 인간 지능을 능가하는 시점에 재평가받는 게 맞을 듯하다. 6–2) 진짜 수확가속 하는가? 저자는 자신의 주장에 법칙을 붙여, 수확가속법칙이라고 말하고 있다. 그런데 과연? Computing power도 저장장치도 모두 기하급수적으로 발전하고 있다고 한다. 그런데 나는 왜 아이폰 16G 폰을 쓰고 있으며, 매번 사진을 지워야 하는 걸까? 수확가속법칙? 나와는 거리가 먼 듯하다. 6–3) AI가 딥러닝 하면 정말 밥까지 떠 먹여주나? AI가 최근에 영화에서 전지전능한 모습을 많이 보여줬고, 알파고가 이세돌을 꺾으면서 AI대세론에도 한층 무게가 실린 것 같다. 하지만 AI에게 바둑을 가르치는 것과 화장지 공장을 운영시키는 것은 매우 다른 문제일 것이다. 각각의 영역에서 AI를 구현하기 위해서는 엄청난 노가다가 선행되어야 하지 않을까? 결국 AI가 딥러닝을 해도 당분간 우리 같은 월급쟁이들이 해야 할 일들은 계속 산더미일 것으로 예상된다. 6–4) 복잡한 인간의 마음을 초당 연산 능력으로 너무 단순화한 듯. 최신형 CPU를 탑재한 컴퓨터에게 렉이 안 걸리는 것을 바랄 뿐, 끊김 없이 말동무가 되어주길 기대하지 않는다. 슈퍼컴퓨터라도 연산 속도만으로 오랜 친구처럼 사람의 미묘한 감정을 헤아리고 반응해줄 수 있을 것 같지는 않다. 빠른 것과 잘하는 것은 다르다. 인간의 마음에는, 통계적 분석 후 최적의 정답을 빠르게 찾는 것 이상의 무엇인가 있는 것 같다. 역시 미묘한 감정을 헤아려주지는 못한다. 6–5) 기억장치가 뇌에만 존재하지 않을 수도.. 저자는 신피질 패턴인식기 개수를 기준으로 인간의 메모리가 20기가 바이트에 불과하다고 한다. 하지만 이런 추정은 기억장치가 뇌에만 존재한다는 가정에 기반한다. 최근 장기 기증한 사람의 성격이나 습관이 수혜자에게 전이되는 현상인 세포 기억설이 조금씩 대두되고 있다. 정상과학으로써 뇌가 모든 기억 및 사고의 역할을 수행한다는 패러다임이 틀릴 수도 있으며, 그럴 경우 신피질을 복제한 인공지능은 인간 뇌와 치환되지 않을 것이다. 세포 기억설 - 위키백과, 우리 모두의 백과사전 문서 수정을 도와주세요. 내용에 대한 의견이 있으시다면 토론 문서에서 나누어 주세요. (2015년 12월 1일)ko.wikipedia.org 6–6) 미지의 영역을 탐구하려는 의지에 박수 인간의 뇌, 의식, 자유의지, 정체성과 같이 건드릴 엄두가 안나는 이슈에 대해서도 누군가는 선구자가 되어 장님 코끼리 만지기 식으로라도 밑그림을 그려야 할 것이다. 어려운 주제에 대해서 논리적인 접근으로 사고의 지평을 넓혀준 저자에게 감사한다. 7. Commitment 7–1) <특이점이 온다>도 빠른 시일 내에 읽어보자! -끝-
[Book Review] 마음의 탄생
1
book-review-마음의-탄생-13aa46bf8387
2018-04-30
2018-04-30 20:42:16
https://medium.com/s/story/book-review-마음의-탄생-13aa46bf8387
false
1,104
null
null
null
null
null
null
null
null
null
Artificial Intelligence
artificial-intelligence
Artificial Intelligence
66,154
Wooho Jung
Engineer, Reader, Imaginer
4de91ff7b744
woohojung
23
19
20,181,104
null
null
null
null
null
null