Your Career Platform for Big Data

Be part of the digital revolution in Germany

 

Latest job opportunities

Aspire Data Recruitment Berlin, Germany
17/07/2018
Full time
Job Title:                           Data Scientist                   Location:                           Berlin Salary:                               €50,000 - €70,000 plus bonus and benefits Our client, the MENA region’s leading driver booking service requires a Data Scientist to join their rapidly developing analytics team based in Berlin. The Data Scientist will: Will work closely with the product and engineering teams to define and confront a range of problems and embark on exploratory data analysis projects to achieve better understanding of phenomena as well as to discover untapped areas of growth and optimization. Help define and track the appropriate key metrics for specific projects. Design and run randomized controlled experiments, analyse the resulting data and communicate results with other teams. Collaborate with engineers to build prototype predictive models into production and re-iterate as needed. Focus on dynamic pricing optimization, optimal dispatching, fraud detection and prevention, ETA prediction and location search optimization. The Data Scientist will: Be innovative, detail-oriented, result-focused, have a solid quantitative background - Advanced degree in a quantitative discipline such as Physics, Statistics, Mathematics, Engineering or Computer Science. Have 2+ years’ experience in data mining, predictive modelling, time series analysis, machine learning, Big Data methodologies, transformation and cleaning of both structured and unstructured data. Be fluent in English along with excellent oral and written communication skills. Be proficient and have demonstrated experience in at least 2 of the following: SQL, R, Python, Spark, Hive, RapidMiner. Have demonstrated experience with database technologies (e.g. Hadoop, Amazon EMR, Hive, Oracle, SAP, DB2, Teradata, MS SQL Server, MySQL) is a plus. Have demonstrated experience with business intelligence and visualization tools (Tableau, MicroStrategy, ChartIO, Qlik) along with geospatial data processing skills is also a plus. Be business savvy, and take full ownership of their work from the inception of the idea to the implementation of the final product. Have knowledge of Agile methodologies  To apply please send your CV or call 01706 825 199
Zühlke Hamburg, Germany
17/07/2018
Full time
Welche Aufgaben erwarten Dich? In Deiner beratenden Funktion begleitest Du unsere Kunden aus dem Bereich Maschinen und Anlagenbau bei der Entwicklung neuer Datenplattformen und eröffnest so den Weg in ein datengetriebenes Geschäftsmodell. Dabei wirkst Du als Experte auch in der Akquise von Projekten mit und bist Teil eines interdisziplinären Bid Teams. Du bist in Deinem Spezialgebiet zentraler Ansprechpartner und konzipierst Lösungen zusammen mit einem interdisziplinären Team von Domänen-Experten und Data-Engineers. Dabei übernimmst Du auch die fachliche Führung sowie die Weiterentwicklung der Kollegen. Deine Tätigkeit als Experte bei Zühlke ist vielfältig – und reicht von der Datenexploration, Aufbereitung von Daten und Feature Engineering bis zur Auswahl geeigneter Machine Learning Algorithmen oder Verfahren. Weil Du eng mit unseren Kunden zusammenarbeitest, können auch Reisen innerhalb Deutschlands Teil Deiner Tätigkeit sein. Was solltest Du mitbringen? Du hast ein technisches Studium abgeschlossen, idealerweise Informatik oder Ingenieurswissenschaften und konntest bereits einschlägige Erfahrung in der Projektarbeit und bei der Auswahl von State-of-the-Art Analyseverfahren sammeln. Projekterfahrung im Bereich IoT, Anomalie Erkennung, Predictive Analytics oder Computer Vision bringst Du ebenso mit. Du verfügst über Programmiererfahrung mit Rapid Prototyping Tools wie z.B. R und/oder Python (Pandas, SciKit, Microsoft Cognitive Toolkit). Du bist erfahren in der Nutzung von Tools zur Datenvisualisierung und Gestaltung von Dashboard Konzepten. Neben fortgeschrittenen Kenntnissen in Statistik und Zeitreihenanalytik verfügst Du auch über Kenntnisse in mindestens zwei der folgenden Methoden: Maschine Learning, Mustererkennung, Modellierung, Regression, Clustering, Classification. Als kundenorientierter Teamplayer, der sich für Innovation und Agilität begeistert sowie sehr gute Deutsch- und Englischkenntnisse mitbringt, passt Du perfekt zu uns. Unser Angebot:  Zühlke unterstützt Deinen Erfolg mit einer einzigartigen Unternehmenskultur, für die wir als einer der besten Arbeitgeber Deutschlands ausgezeichnet wurden. Es ist uns wichtig, dass wir Dich durch kontinuierliche, hochwertige Weiterbildungen fördern. Wir sind interdisziplinär ausgerichtet, tauschen uns regelmäßig aus und lernen voneinander - hier spürst Du, dass wir offen miteinander kommunizieren, uns selbst ehrlich einschätzen und gerne im Team arbeiten. Dazu bieten wir ein attraktives Leistungspaket, moderne Arbeitsplätze und die gemeinsame Leidenschaft, jeden Tag das Beste zu geben!  Freue Dich außerdem auf die kollegiale Zusammenarbeit in einem motivierten und dynamischen Team vielfältige Angebote für Deine Gesundheit und Fitness einen Arbeitsplatz, an dem eine Vertrauenskultur gelebt wird und Du viel Gestaltungsfreiheit besitzt Lass Dich vom Zühlke Spirit anstecken und lies unsere  Success Stories  aus den Projekten
Venturi Hamburg, Germany
17/07/2018
Full time
Hadoop ist für Sie kein Wort, das nach Comic Sprechblase klingt? Sie möchten eine herausfordende Tätigkeit in einem gestandenen Deutschen Unternehmen wo Sie noch viel mitgestalten können? Dann bin ich auf Sie gespannt! Ihr Profil: Ein abgeschlossenes Studium mit einem Analytischen Schwerpunkt/Wirtschaftsinformatik Mindestens 1 Jahr aktive Hadoop sowie Apache Spark/Hive Erfahrung Data Mining lässt Sie weniger an dunkle Gänge und mehr an spannende Projekte denken Mindestens 1 Jahr Erfahrung im Umfeld der Datenanalyse, Big Data, Machine Learning Gute Python, Spark, Cloudera und R Kontakt von Vorteil Erster Kontakt mit weiteren DWH und BI Lösungen von Vorteil; Erfahrungen mit ETL Begeisterung für neue und spannende Technologien Ihre Aufgaben: Sie analysieren aktuelle Fachbereichanforderungen und setzen diese in Lösungen um Sie Entwickeln eine neue Big Data Plattform und arbeiten eng mit Data Scientisten zusammen Sie übernehmen Verantwortung für die Aufbereitung großer Datenmengen Sie arbeiten eng mit Ihre Big Data Kollegen zusammen Sie bringen Ihre Erfahrung und Ideen in das gesamte BI/Big Data Team ein und sind bereit Neues zu lernen Festanstellung: ja  Gehaltsmöglichkeiten: bis 85.000 Euro p.a. je nach Einstiegslevel  Reisetätigkeit: Nein, Inhouse  Beginn: ASAP  Unbefristet: ja Kontakt: Charlotte Sommerfeldt  Email: charlotte.sommerfeldt@venturi-group.com Wenn Sie auf der Suche nach einer neuen Herausforderung im Bereich Business Intelligence sind und gerne Teil eines hoch motivierten Teams werden möchten, dass Ihre Ideen und Erfahrungen schätzt, dann sollten Sie sich auf diese Stelle bewerben.  Für nähere Informationen zu dieser Stelle und weitere Möglichkeiten für Sie im BI Umfeld kommen Sie gerne direkt auf mich per Mail oder Telefon zu: charlotte.sommerfeldt@venturi-group.com / +49800 6644810 Ich freue mich auf Ihre Kontaktaufnahme.  Venturi ist ein international agierendes Unternehmen und spezialisiert auf die Personalberatung und – vermittlung im Bereich Business Intelligence (BI), Development & Design, Legal IT, Support & Infrastructure, Testing, Senior Appointments.  Ich arbeite serviceorientiert für Kandidaten- und Kundenseite und möchte immer den höchst möglichen Standard und Qualität in meiner Arbeit bieten. Mir ist es wichtig offene Positionen mit geeigneten Kandidaten zu besetzen und langfristig Partnerschaften mit Kandidaten und Unternehmen aufzubauen. Germany ,  Hamburg
Venturi Hamburg, Germany
17/07/2018
Full time
Für meinen Kunden mit internationaler Präsenz und Hauptsitz in Hamburg suche ich aktuell einen Senior Data Scientist (m/w), der das fünf-Köpfige Data Science Team InHouse unterstützen möchte. Wenn Sie also Ihre Analytics Skills, Python und R Erfahrungen und SQL Kenntnisse erfolgreich in einem neuen Projekt einsetzen möchten, dann wäre dies eine einmalige Gelegenheit. Ihre Aufgaben: Kommunikation mit den Fachbereichen, um effiziente Lösungen zu entwickeln Entwicklung von analytischen Lösungen, die auf Prognosen beruhen Konzeption von Tests zur Weiterentwicklung von internen Prozessen Vorwärtsdenken, um das Unternehmen nach vorne zu bringen Anforderungsentwicklung für die technischen Systeme Ihre Erfahrung: Mind. 3-5 Jahre Berufserfahrung im Bereich Data-Mining, Machine Learning und Modellentwicklung Praxis- Erfahrung in der Simulation von Modellanwendungen Erfahrung in der Konzeption und Umsetzung von analytischen Modellen in technische Systeme Hands-on Erfahrung mit Data-Mining und Statistik-Tools sowie SQL Kommunikationsstark und hohes Durchsetzungsvermögen Hohe Innovationsdenken und begeisterungsfähig Region: Hamburg Unbefristet: ja Festanstellung: ja Reisebereitschaft: Nein, InHouse Stelle Start: ASAP Gehalt: je nach Profil Wenn Sie auf der Suche nach einer neuen Herausforderung im Bereich Business Intelligence sind und gerne Teil eines hoch motivierten Teams werden möchten, dass Ihre Ideen und Erfahrungen schätzt, dann sollten Sie sich auf diese Stelle Bewerben. Für nähere Informationen zu dieser Stelle und weitere Möglichkeiten für Sie im BI Umfeld kommen Sie gerne direkt auf mich per Mail oder Telefon zu: charlotte.sommerfeldt@venturi-group.com  / 0800 6644810 Ich freue mich auf Ihre Kontaktaufnahme. Venturi ist ein international agierendes Unternehmen und spezialisiert auf die Personalberatung und – vermittlung im Bereich Business Intelligence (BI), Development & Design, Legal IT, Support & Infrastructure, Testing, Senior Appointments. Ich arbeite serviceorientiert für Kandidaten- und Kundenseite und möchte immer den höchst möglichen Standard und Qualität in meiner Arbeit bieten. Mir ist es wichtig offene Positionen mit geeigneten Kandidaten zu besetzen und langfristig Partnerschaften mit Kandidaten und Unternehmen aufzubauen. Germany ,  Hamburg  

DataCareer Blog

Companies use machine learning to improve their business decisions. Algorithms select ads, predict consumers’ interest or optimize the use of storage. However, few stories of machine learning applications for public policy are out there, even though public employees often make comparable decisions. Similar to the business examples, decisions by public employees often try to optimize the use of limited resources. Algorithms may assist tax authorities in improving the allocation of available working hours, or help bankers make lending decisions. Similarly, algorithms can be employed to guide decisions taken by social workers or judges. // This blogpost lists three research papers that analyze and discuss the use of machine learning for very specific problems in public policy. While the potential seems huge, we do not want to neglect some of the many potential pitfalls for machine learning in public policy. Business applications often maximize profits. For policy decisions, however, the maximizable outcome may be harder to define or multidimensional. In many cases, not all relevant outcome dimensions are directly observable and measurable, which makes it more difficult to evaluate the impact of an algorithm. Tech companies would usually obtain training datasets through experimenting, while datasets for public policy often contain only one outcome for a specific group of people. If tax authorities never scrutinize restaurants, how can we form a predictive model for this industry? Predictions for public policy problems often face this so-called selected labels problem and it needs innovative approaches and the willingness to perform randomized experiments to get around it. This is just a brief list. Susan Athey’s paper provides more food for thought on the potential - and potential pitfalls - of using prediction in public policy.   Research on Machine Learning Applications in Public Policy Improving refugee integration through data-driven algorithmic assignment Developed democracies are settling an increased number of refugees, many of whom face challenges integrating into host societies. We developed a flexible data-driven algorithm that assigns refugees across resettlement locations to improve integration outcomes. The algorithm uses a combination of supervised machine learning and optimal matching to discover and leverage synergies between refugee characteristics and resettlement sites. The algorithm was tested on historical registry data from two countries with different assignment regimes and refugee populations, the United States and Switzerland. Our approach led to gains of roughly 40 to 70%, on average, in refugees’ employment outcomes relative to current assignment practices. This approach can provide governments with a practical and cost-efficient policy tool that can be immediately implemented within existing institutional structures. Bansak, K., Ferwerda, J., Hainmueller, J., Dillon, A., Hangartner, D., Lawrence, D., & Weinstein, J.; Science, 2018 Switzerland is currently implementing an algorithm based allocation of refugees. We are excited to see first results!   Human Decisions and Machine Predictions Can machine learning improve human decision making? Bail decisions provide a good test case. Millions of times each year, judges make jail-or-release decisions that hinge on a prediction of what a defendant would do if released. The concreteness of the prediction task combined with the volume of data available makes this a promising machine-learning application. Yet comparing the algorithm to judges proves complicated. First, the available data are generated by prior judge decisions. We only observe crime outcomes for released defendants, not for those judges detained. This makes it hard to evaluate counterfactual decision rules based on algorithmic predictions. Second, judges may have a broader set of preferences than the variable the algorithm predicts; for instance, judges may care specifically about violent crimes or about racial inequities. We deal with these problems using different econometric strategies, such as quasi-random assignment of cases to judges. Even accounting for these concerns, our results suggest potentially large welfare gains: one policy simulation shows crime reductions up to 24.7% with no change in jailing rates, or jailing rate reductions up to 41.9% with no increase in crime rates. Moreover, all categories of crime, including violent crimes, show reductions; these gains can be achieved while simultaneously reducing racial disparities. These results suggest that while machine learning can be valuable, realizing this value requires integrating these tools into an economic framework: being clear about the link between predictions and decisions; specifying the scope of payoff functions; and constructing unbiased decision counterfactuals. Jon Kleinberg  Himabindu Lakkaraju  Jure Leskovec Jens Ludwig  Sendhil Mullainathan; Quarterly Journal of Economics, 2018 // Using Text Analysis to Target Government Inspections: Evidence from Restaurant Hygiene Inspections and Online Reviews Restaurant hygiene inspections are often cited as a success story of public disclosure. Hygiene grades influence customer decisions and serve as an accountability system for restaurants. However, cities (which are responsible for inspections) have limited resources to dispatch inspectors, which in turn limits the number of inspections that can be performed. We argue that NLP can be used to improve the effectiveness of inspections by allowing cities to target restaurants that are most likely to have a hygiene violation. In this work, we report the first empirical study demonstrating the utility of review analysis for predicting restaurant inspection results. Kang, J. S., Kuznetsova, P., Choi, Y., Luca, M., 2013 , Technical Report Here is related paper on the same topic suggesting ways for governments on how to obtain the required expertise: Crowdsourcing City Government: Using Tournaments to Improve Inspection Accuracy Further readings: Two papers with an excellent overview on the topic Machine Learning: An Applied Econometric Approach Prediction Policy Problems The Economist on the same topic: Of prediction and policy, The Economist 2016  
Are you looking for real world data science problems to sharpen your skills? In this post, we introduce you to four platforms hosting data science competitions. Data science competitions can be a great way for gaining practical experience with real world data, and for boosting your motivation through the competitive environment they provide. Check them out, competitions are a lot of fun! Kaggle Kaggle is the best known platform for data science competitions. Data scientists and statisticians compete to create the best models for describing and predicting the data sets uploaded by companies or NGOs. From predicting house prices in the US to demographics of mobile phone users in China or the properties of soil in Africa, Kaggle offers many interesting challenges to solve real world problems. Check out their No Free Hunch Blog featuring the winners of each competition. The platform was recently acquired by Alphabet, Google’s parent company, and also offers a wide range of datasets to train your algorithms and other useful resources to improve your data science skill set.   // DrivenData Similar to other platforms, the dataset is available online and participants submit their best predictive models. The great thing about DrivenData competitions is that the competition question and datasets are related to the work of non-profits, which can be especially interesting to those who want to contribute to a good cause. Furthermore, the data problems are no less diverse and range from predicting dengue fever cases, to estimating the penguin population in the Antarctic and forecasting energy consumption levels.  For some challenges, the best model wins a prize, for others you get the glory and the knowledge that you applied your skillset to make the world a better place. DrivenData offers great opportunities to tackle real-world problems with real-world impact. Numerai Numerai is a data science competition platform focusing on finance applications. What makes their competitions particularly interesting is that the participants’ predictions are used in the underlying hedge fund. Data scientists entering Numerai’s tournaments currently receive an encrypted data set every week. The data set is an abstract representation of stock market information that preserves its structure without revealing details. The data scientists then create machine-learning algorithms to find patterns in the data, and they test their models by uploading their predictions to the website. Numerai, then creates a meta-model from all submissions to make its investments. The models get ranked, with the top 100 earning Numeraire coins, a cryptocurrency launched by Numerai. Numerai's mix of data science, cryptography, artificial intelligence, crowdsourcing and bitcoin has given the fledgling business an exciting flair.   Tianchi Tianchi is a data competition platform by Alibaba Cloud, the cloud computing arm of Alibaba Group, and has strong similarities with Kaggle. The platform focuses on Chinese data scientist, but most pages are also available in English. Tianchi boasts a community of over 150,000 data scientists, 3,000 institutes and business groups from over 80 countries. Besides the competitions, the platform also offers datasets and a notebook to run Python 3 scripts.       //
Curious about neural networks and deep learning? This post will inspire you to get started in deep learning. Why are we witnessing this kind of build up for neural networks? It is because of their amazing applications. Some of their applications include image classification, face recognition, pattern recognition, automatic machine translation, and so on. So, let’s get started now. Machine Learning is a field of computer science that provides computers the capability to learn and improve from experience without being programmed explicitly. Deep learning is a form of machine learning that uses a computing model that is highly inspired by the structure of the brain. Hence, we call this computing model as a Neural Network. A neural network is a computing system comprising highly interconnected and simple processing elements which process the information through their dynamic state response to external inputs. A ‘neuron’ is the fundamental processing element of a neural network. The neural network comprises a large number of neurons working simultaneously to solve specific problems. This article explains the concept of neural networks and why they are a vital component in the process of deep learning. It also helps to let you know:- The advantages of neural networks over conventional techniques Working of Neural networks, Working of a Neural Network - Training, Working of a Neural Network - Learning Rules Network models and algorithms of Neural Networks   Why Neural Networks Matter in Deep Learning? Consider machine learning as a pack horse for processing information, then a carrot that draws the horse forward is the neural network. A system should not be programmed to execute a specific task for it to be able to learn truly; instead, it must be programmed to learn to execute the task. To accomplish this, the system uses deep learning (a more refined form of machine learning) which is based on neural networks. With the help of neural networks, the system can perceive data patterns independently to learn how to execute a task.   Advantages of Neural Networks over Conventional Techniques Depending on the strength of internal data patterns and the nature of the application, you can usually expect a network to train well. This is applied to problems where the relationships may be quite nonlinear or dynamic. Very often, the conventional techniques are limited by strict assumptions of variable independence, linearity, normality, etc. As neural network can capture various types of relationships, it enables the user to relatively easily and quickly model phenomena which otherwise may have been impossible or very difficult to explain.     Working of a Neural Network Neural networks are modeled after the neuronal structure of the brain’s cerebral cortex but on smaller scales. They are usually organized in layers. Layers are comprised of many nodes which are interconnected and contain an activation function. The patterns are presented to the network through the input layer. This layer communicates to hidden layers (one or more in number) where the real processing is carried out through a system of weighted connections. Then, the hidden layers(neural hidden layer as shown in the below figure) are connected to an output layer(neural output layer as shown in the below figure) and it is the answer as depicted in the image shown below. The information flows via a neural network in 2 ways. When the neural network is operating normally (after its training) or learning (during training), the information patterns are fed into the network through input units. These input units will trigger the hidden unit layers and these in turn will arrive at the output units. This design is considered as the feedforward network. Every unit gets inputs from the units situated on its left. Then, the inputs are multiplied by the connections’ weights they travel along. Each unit sums up every input it receives in its way and the unit triggers the units situated on its right if the sum is more than a certain threshold value. In the below section, we will see how a neural network learns.   Working of a Neural Network - Training Training a neuron involves applying a set of steps to adjust the thresholds and weights of its neurons. This kind of adjustment process (also known as learning algorithm) tunes the network so that the outputs of the network are very close to the desired values. The network is ready to be trained once it is structured for a specific application. The initial weights are selected randomly to begin this process. Then, the training or learning starts. There are two approaches to training - unsupervised and supervised. In supervised training, the network is provided with the desired output in two ways. The first one involves manually grading the performance of the network and the second one is by allocating the desired outputs with the inputs. In unsupervised training, the network must make sense of the inputs without the help from outside. To put this in familiar terms, let’s consider an instance. Your kids are called supervised if you provide a solution to them during every situation in their life. They are called unsupervised if your kids make decisions on their own out of their understanding.   Most of the neural networks consist of some form of learning rule which alters the weights of connections according to the input patterns that are presented to it. Like their biological counterparts, the neural networks learn by example.     Working of a Neural Network - Learning Rules Neural networks use various kinds of learning rules. They are as follows. Hebbian Learning Rule - This learning rule determines, how to alter the weight of nodes of a network. Perceptron Learning Rule - The network begins its learning by allocating a random value to each weight. Delta Learning Rule - The modification in a node’s sympatric weight is equal to the multiplication of input and the error. Correlation Learning Rule - It is the supervised learning. Outstar Learning Rule - It can be used when it assumes that neurons or nodes in a network are arranged in a layer. The Delta Learning Rule is often used by the most common class of neural networks known as BPNNs (backpropagation neural networks). Backpropagation implies the backward propagation of error.   // Major Neural Network Models The primary neural network models are as follows. Multilayer perceptron - This neural network model maps the input data sets onto a set of appropriate outputs. Radial Basis Function Network - This neural network uses radial basis functions as activation functions. Both the above models are supervised learning networks, and they are used with one or more dependent variables at the output. Kohonen Network - This is an unsupervised learning network. This is used for clustering process.   Neural Network Algorithms As I stated earlier, the procedure used to perform the learning process in a neural network is known as the training algorithm. There are various training algorithms with different performance and characteristics. The major ones are Gradient Descent (used to find the function’s local minimum) and Evolutionary Algorithms (based on the concept of survival of the fittest or natural selection in biology).   Deep Neural Networks Deep Neural Networks can be thought of as the components of broader applications of machine learning that involve algorithms for regression, classification, and reinforcement learning(a goal-oriented learning depending on interaction with the environment). These networks are distinguished from single-hidden-layer neural networks by their depth. This implies the number of node layers through which the data passes in a pattern recognition’s multi-step process. Conventional machine learning depends on shallow networks that are composed of one output and one input layer with at most one hidden layer in-between. Including input and the output, more than three layers qualify as ‘deep’ learning. A deep neural network is shown in the below figure which has three hidden layers apart from the input and output layers. Hence, deep is a technical and strictly defined term that implies more than one hidden layer. Based on the previous layer’s output, each layer of nodes trains on a different feature set in deep neural networks.   Unlike most traditional machine learning algorithms, deep neural networks carry out automatic feature extraction without intervention. These networks can discover latent structures within unstructured(raw data), unlabeled data which is the majority of data in the world. A deep neural network which is trained on labeled data can be applied to raw data. This gives the deep neural network access to much more input when compared with machine learning networks. This indicates higher performance as the accuracy of a network depends on how much data it is trained on. Training on more data results in higher accuracy.   Applications of Neural Networks in Python and R Python Libraries using Neural Networks   Theano Theano is an open source project released under the BSD license. At its heart, Theano is a compiler for mathematical expressions in Python. It knows how to take your structures and turn them into very efficient code that uses NumPy, efficient native libraries like BLAS and native code (C++) to run as fast as possible on CPUs or GPUs. It uses a host of clever code optimizations to squeeze as much performance as possible from your hardware. The actual syntax of Theano expressions is symbolic, which can be off putting to beginners used to normal software development. Specifically, expression are defined in the abstract sense, compiled and later actually used to make calculations. It was specifically designed to handle the types of computation required for large neural network algorithms used in Deep Learning. It was one of the first libraries of its kind and is considered an industry standard for Deep Learning research and development.   TensorFlow TensorFlow is an open source library for fast numerical computing. It was created and is maintained by Google and released under the Apache 2.0 open source license. The API is nominally for the Python programming language, although there is access to the underlying C++ API. Unlike other numerical libraries intended for use in Deep Learning like Theano, TensorFlow was designed for use both in research and development and in production systems, not least RankBrain in Google search and the fun Deep Dream project. It can run on single CPU systems, GPUs as well as mobile devices and large scale distributed systems of hundreds of machines. It’s easy to classify TensorFlow as a neural network library, but it’s not just that. Yes, it was designed to be a powerful neural network library. But it has the power to do much more than that. You can build other machine learning algorithms on it such as decision trees or k-Nearest Neighbors. You can literally do everything you normally would do in numpy! It’s aptly called “numpy on steroids.”   R Libraries using Neural Networks   Caret The caret package is a set of tools for building machine learning models in R. The name “caret” stands for C lassification A nd RE gression T raining. As the name implies, the caret package gives you a toolkit for building classification models and regression models. Moreover, caret provides you with essential tools for data splitting, pre-processing, feature selection, model tuning using resampling, variable importance estimation as well as other functionality. There are many different modeling functions in R. Some have different syntax for model training and/or prediction. The package started off as a way to provide a uniform interface the functions themselves, as well as a way to standardize common tasks (such parameter tuning and variable importance). Caret provides a simple, common interface to almost every machine learning algorithm in R. When using caret, different learning methods like linear regression, neural networks, and support vector machines, all share a common syntax (the syntax is basically identical, except for a few minor changes). Moreover, additional parts of the machine learning workflow – like cross validation and parameter tuning – are built directly into this common interface. To say that more simply, caret provides you with an easy-to-use toolkit for building many different model types and executing critical parts of the ML workflow. This simple interface enables rapid, iterative modeling. In turn, this iterative workflow will allow you to develop good models faster, with less effort, and with less frustration.   nnet There are many ways to create a neural network. You can code your own from scratch using a programming language such as C# or R. You can also use a tool such as the open source Weka or Microsoft Azure Machine Learning. The R language has an add-on package named nnet that allows you to create a neural network classifier. The nnet R package has been created by Brian Ripley. You can evaluate the accuracy of the model and make predictions using the nnet package. The functions in the nnet package allow you to develop and validate the most common type of neural network model, i.e, the feed-forward multi-layer perceptron. The functions have enough flexibility to allow the user to develop the best or most optimal models by varying parameters during the training process.   Conclusion Neural networks have broad applicability to business problems in the real world. They are currently used applied in various industries, and their applicability is getting increased day-by-day. The primary neural network applications include stock exchange prediction, image compression, handwriting recognition, fingerprint recognition, feature extraction, and so on. But, there is a lot more research that is going on in neural networks.   Author: Savaram Ravindra is a writer on Mindmajix.com working on data science related topics. Previously, he was a Programmer Analyst at Cognizant Technology Solutions. He holds a MS degree in Nanotechnology from VIT University
View all blog posts


Looking for data professionals?

 

Post a Job