Back to all Blog Posts

The Rosenblatt Perceptron – The Early Beginnings of Deep Learning

  • Deep Learning
  • Python
08. December 2017
·

Fabian Müller
COO

The Perceptron was the first type of artificial neuron and was first introduced by Frank Rosenblatt in the late 1950s. Its design was inspired by the McCulloch-Pitts neuron model. While other types of neurons have since replaced the Perceptron, its fundamental design is still applied in modern neural networks.

The Perceptron can be used for learning linearly separable classifications. It processes inputs$ \left[ x_{1}, x_{2}, ... , x_{n} \right] $ into a binary output $ y_{i} $.Weights $ \left[ w_{1}, w_{2}, ... , w_{n} \right] $ determine the importance of each input for the final output. The output is calculated as the weighted sum of the inputs:

$y_{i }=\sum_{i}{w _{i} x_{i}}$

Schematic Representation of the Perceptron

To ensure that $y_{ i }$ is a binary outcome, the Perceptron uses a step function (also called a hard limiter) with an estimated threshold, which is also referred to as bias (represented as the unit input in the diagram above):

$y_{i} = \left\{ \begin{array}{l} 0 \quad \text{if } w \times x + b \lt 0 \\ 1 \quad \text{otherwise} \end{array} \right.$

Here, $ w \times x \equiv \sum_{i}{{w_i} x_{i}} $ represents the dot product of $w$ and $x$ with the bias $b$. A step function is a non-linear function that maps the weighted sum to the desired output range. Side note: Even modern neural networks require a non-linear function—however, compared to the step function, their form is smoother (which is why they are often called soft limiters).

Example of a Step Function

The Perceptron learns by iteratively adjusting the weight vector $w$. This process is defined as follows:

$w \gets \dot{w} + v \times (y_{i} - \hat{y}_{i}) \times x_{i}$

Here, $\dot{w}$ represents the previous weight vector, $v \in (0, \infty)$ is the learning rate, and $(y_{i} - \hat{y}_{i})$ is the error in the current iteration. The weight adjustment is thus computed based on the previous error, scaled by the learning rate, and weighted by the input. This is known as the Perceptron learning rule.

The following Python code illustrates the Perceptron learning mechanism for the simple problem below:

Given are the data points $x_{i} = \left\{ \begin{bmatrix} x_{1}, x_{2} \end{bmatrix} \right\}$ and the vector $y_{i}$ of the associated outputs. In addition, $y_{i} = 1$ if $x_{1} = 1$ or $x_{2} = 1$ and $y_{i} = 0$ otherwise.

# Coden des Rosenblatt Perzeptrons  
# ---------------------------------------------------------- 
import numpy as np 
import random 
 
random.seed(1) 

# Treppenfunktion 
def unit_step(x): 
    if x  {} | {}".format('Index', 'Probability', 'Prediction', 'Target')) 
print('---') 

for index, x in enumerate(X): 
    y_hat = np.dot(x, w) 
    print("{}: {} -> {} | {}".format(index, round(y_hat, 3),
                                     unit_step(y_hat), y[index])) 
%matplotlib inline 
import matplotlib.pyplot as plt 

# Grafik Trainingsfehler  
plt.plot(range(n),errors) 
plt.xlabel('Epoche') 
plt.ylabel('Fehler') 
plt.title('Trainingsfehler') 
plt.show() 

Development of Training Error Over Training Iterations (Epochs)

Inputs, Predicted Probabilities, and Outputs

Index Probability Forecast Target
0 -0.033 0 0
1 0.382 1 1
2 0.521 1 1
3 0.936 1 1

After several hundred iterations, the Perceptron has adjusted the weights such that all data points are correctly predicted. This demonstrates that the Perceptron is capable of solving our simple classification problem. As seen in the graph, 100 epochs are more than sufficient, and we could have stopped the learning process earlier.

It is important to note that this is a pure in-sample classification. The notorious memorization of training data (overfitting) is therefore not an issue. In real-world problems, overfitting is typically prevented by stopping the learning process earlier—this is known as early stopping.

Limitations and Constraints of the Perceptron

Shortly after its publication in the early 1960s, the Perceptron gained significant attention and was widely regarded as a powerful learning algorithm. However, this perception changed dramatically following the famous critique by Minsky and Papert (1969) in the late 1960s and early 1970s.

Minsky and Papert proved that the Perceptron is highly limited in what it can learn. Specifically, they demonstrated that the Perceptron requires the correct features to successfully learn a classification task. With sufficient hand-selected features, the Perceptron performs very well. However, without carefully chosen features, it immediately loses its learning ability.

Yudkowsky (2008, p. 15f.)(3) describes an example of the Perceptron's failure from the early days of neural networks:

„Once upon a time, the US Army wanted to use neural networks to automatically detect camouflaged enemy tanks. The researchers trained a neural net on 50 photos of camouflaged tanks in trees, and 50 photos of trees without tanks. Using standard techniques for supervised learning, the researchers trained the neural network to a weighting that correctly loaded the training set—output “yes” for the 50 photos of camouflaged tanks, and output “no” for the 50 photos of forest. This did not ensure, or even imply, that new examples would be classified correctly."

The neural network might have “learned” 100 special cases that would not generalize to any new problem. Wisely, the researchers had originally taken 200 photos, 100 photos of tanks and 100 photos of trees. They had used only 50 of each for the training set. The researchers ran the neural network on the remaining 100 photos, and without further training the neural network classified all remaining photos correctly. Success confirmed! The researchers handed the finished work to the Pentagon, which soon handed it back, complaining that in their own tests the neural network did no better than chance at discriminating photos.

It turned out that in the researchers’ dataset, photos of camouflaged tanks had been taken on cloudy days, while photos of plain forest had been taken on sunny days. The neural network had learned to distinguish cloudy days from sunny days, instead of distinguishing camouflaged tanks from empty forest.

Our previous example, in which the perceptron successfully learnt the input $X= \left\{ \left[ 0,0 \right], \left[ 0,1 \right], \left[ 1,0 \right], \left[ 1,1 \right] \right\} $, to correctly classify the output $ y = \{ 0, 1, 1, 1 \} $ can be understood as a logical OR function $ (y_{i} = 1 \text{ if } x_{1} = 1 \text{ or } x_{2} = 1) $. In this problem, OR is defined as non-exclusive.

The problem can be converted into an exclusive XOR function by the following modification: $ y = \{ 0, 1, 1, 0 \} $. The question is whether the perceptron is still able to approximate this new function correctly?

random.seed(1)

# Dieselben Daten 
X = np.array([[0,0],  
              [0,1],  
              [1,0],  
              [1,1]]) 

# Zurücksetzen der Vektoren für Gewichte und Fehler 
w = np.random.rand(2) 
errors = [] 

# Aktualisieren der Outputs 
y = np.array([0,1,1,0]) 

# Nochmals: Training ... 
for i in range(n): 
    # Zeilenindex 
    index = random.randint(0,3) 

    # Minibatch 
    x_batch = X[index,:] 
    y_batch = y[index] 

    # Aktivierung berechnen 
    y_hat = unit_step(np.dot(w, x_batch)) 

    # Fehler berechnen und abspeichern 
    error = y_batch - y_hat 
    errors.append(error) 

    # Gewichte anpassen 
    w += eta * error * x_batch 

# ... und Vorhersage   
for index, x in enumerate(X): 
    y_hat = np.dot(x, w) 
    print("{}: {} -> {} | {}".format(index, round(y_hat, 3), unit_step(y_hat), y[index])) 
     
%matplotlib inline 
import matplotlib.pyplot as plt 

# Grafik Trainingsfehler  
plt.plot(range(n),errors)
plt.xlabel('Epoche') 
plt.ylabel('Fehler') 
plt.title('Trainingsfehler') 
plt.show() 

Development of the training error over the training iterations (Epochs)

Inputs, predicted probabilities and outputs

Index Probability Forecast Target
0 0.000 1 0
1 0.024 1 1
2 0.156 1 1
3 0.180 1 1

The peculiarities of the modified problem make it difficult for the perceptron to learn the task correctly, given the selected inputs. As discussed by Minsky and Papert, the features are the key to the perceptron solving the problem.

Conclusion

Unfortunately, Minsky and Papert's critique has been misunderstood by a large part of the scientific community: If learning the right features is essential and neural networks alone are incapable of learning these features, they are useless for all non-trivial learning tasks. This interpretation remained a widely shared consensus for the next 20 years and led to a dramatic reduction in scientific interest in neural networks as learning algorithms.

As a further consequence of Minsky and Papert's findings, other learning algorithms have since taken the place of neural networks. One of the most prominent algorithms was the Support Vector Machine (SVM). By transforming the input space into a non-linear feature space, the SVM solved the problem that apparently caused neural networks to fail.

It was not until the late 1980s and early 1990s that it was slowly realised that neural networks were capable of learning useful, non-linear features from data. This was made possible by stringing together several layers of neurons. Each layer learns from the outputs of the previous one. This allows useful features to be derived from the original input data - i.e. precisely the problem that a single peceptron cannot solve.

Even if today's deep learning (i.e. neural networks with many layers) is significantly more complex than the original perceptron, it is still based on the basic techniques of the perceptron.

References

1. Rosenblatt, F. (1958). The perceptron: A probabilistic model for information storage and organization in the brain. Psychological review, 65(6), 386.2. Minsky, M. L., & Papert, S. A. (1987). Perceptrons-Expanded Edition: An Introduction to Computational Geometry.3. Yudkowsky, E. (2008). Artificial intelligence as a positive and negative factor in global risk. Global catastrophic risks, 1(303), 184.

Linkedin Logo
Marcel Plaschke
Head of Strategy, Sales & Marketing
schedule a consultation
Zugehörige Leistungen
No items found.

More Blog Posts

  • Artificial Intelligence
AI Trends Report 2025: All 16 Trends at a Glance
Tarik Ashry
05. February 2025
Read more
  • Artificial Intelligence
  • Data Science
  • Human-centered AI
Explainable AI in practice: Finding the right method to open the Black Box
Jonas Wacker
15. November 2024
Read more
  • Artificial Intelligence
  • Data Science
  • GenAI
How a CustomGPT Enhances Efficiency and Creativity at hagebau
Tarik Ashry
06. November 2024
Read more
  • Artificial Intelligence
  • Data Culture
  • Data Science
  • Deep Learning
  • GenAI
  • Machine Learning
AI Trends Report 2024: statworx COO Fabian Müller Takes Stock
Tarik Ashry
05. September 2024
Read more
  • Artificial Intelligence
  • Human-centered AI
  • Strategy
The AI Act is here – These are the risk classes you should know
Fabian Müller
05. August 2024
Read more
  • Artificial Intelligence
  • GenAI
  • statworx
Back to the Future: The Story of Generative AI (Episode 4)
Tarik Ashry
31. July 2024
Read more
  • Artificial Intelligence
  • GenAI
  • statworx
Back to the Future: The Story of Generative AI (Episode 3)
Tarik Ashry
24. July 2024
Read more
  • Artificial Intelligence
  • GenAI
  • statworx
Back to the Future: The Story of Generative AI (Episode 2)
Tarik Ashry
04. July 2024
Read more
  • Artificial Intelligence
  • GenAI
  • statworx
Back to the Future: The Story of Generative AI (Episode 1)
Tarik Ashry
10. July 2024
Read more
  • Artificial Intelligence
  • GenAI
  • statworx
Generative AI as a Thinking Machine? A Media Theory Perspective
Tarik Ashry
13. June 2024
Read more
  • Artificial Intelligence
  • GenAI
  • statworx
Custom AI Chatbots: Combining Strong Performance and Rapid Integration
Tarik Ashry
10. April 2024
Read more
  • Artificial Intelligence
  • Data Culture
  • Human-centered AI
How managers can strengthen the data culture in the company
Tarik Ashry
21. February 2024
Read more
  • Artificial Intelligence
  • Data Culture
  • Human-centered AI
AI in the Workplace: How We Turn Skepticism into Confidence
Tarik Ashry
08. February 2024
Read more
  • Artificial Intelligence
  • Data Science
  • GenAI
The Future of Customer Service: Generative AI as a Success Factor
Tarik Ashry
25. October 2023
Read more
  • Artificial Intelligence
  • Data Science
How we developed a chatbot with real knowledge for Microsoft
Isabel Hermes
27. September 2023
Read more
  • Data Science
  • Data Visualization
  • Frontend Solution
Why Frontend Development is Useful in Data Science Applications
Jakob Gepp
30. August 2023
Read more
  • Artificial Intelligence
  • Human-centered AI
  • statworx
the byte - How We Built an AI-Powered Pop-Up Restaurant
Sebastian Heinz
14. June 2023
Read more
  • Artificial Intelligence
  • Recap
  • statworx
Big Data & AI World 2023 Recap
Team statworx
24. May 2023
Read more
  • Data Science
  • Human-centered AI
  • Statistics & Methods
Unlocking the Black Box – 3 Explainable AI Methods to Prepare for the AI Act
Team statworx
17. May 2023
Read more
  • Artificial Intelligence
  • Human-centered AI
  • Strategy
How the AI Act will change the AI industry: Everything you need to know about it now
Team statworx
11. May 2023
Read more
  • Artificial Intelligence
  • Human-centered AI
  • Machine Learning
Gender Representation in AI – Part 2: Automating the Generation of Gender-Neutral Versions of Face Images
Team statworx
03. May 2023
Read more
  • Artificial Intelligence
  • Data Science
  • Statistics & Methods
A first look into our Forecasting Recommender Tool
Team statworx
26. April 2023
Read more
  • Artificial Intelligence
  • Data Science
On Can, Do, and Want – Why Data Culture and Death Metal have a lot in common
David Schlepps
19. April 2023
Read more
  • Artificial Intelligence
  • Human-centered AI
  • Machine Learning
GPT-4 - A categorisation of the most important innovations
Mareike Flögel
17. March 2023
Read more
  • Artificial Intelligence
  • Data Science
  • Strategy
Decoding the secret of Data Culture: These factors truly influence the culture and success of businesses
Team statworx
16. March 2023
Read more
  • Artificial Intelligence
  • Deep Learning
  • Machine Learning
How to create AI-generated avatars using Stable Diffusion and Textual Inversion
Team statworx
08. March 2023
Read more
  • Artificial Intelligence
  • Human-centered AI
  • Strategy
Knowledge Management with NLP: How to easily process emails with AI
Team statworx
02. March 2023
Read more
  • Artificial Intelligence
  • Deep Learning
  • Machine Learning
3 specific use cases of how ChatGPT will revolutionize communication in companies
Ingo Marquart
16. February 2023
Read more
  • Recap
  • statworx
Ho ho ho – Christmas Kitchen Party
Julius Heinz
22. December 2022
Read more
  • Artificial Intelligence
  • Deep Learning
  • Machine Learning
Real-Time Computer Vision: Face Recognition with a Robot
Sarah Sester
30. November 2022
Read more
  • Data Engineering
  • Tutorial
Data Engineering – From Zero to Hero
Thomas Alcock
23. November 2022
Read more
  • Recap
  • statworx
statworx @ UXDX Conf 2022
Markus Berroth
18. November 2022
Read more
  • Artificial Intelligence
  • Machine Learning
  • Tutorial
Paradigm Shift in NLP: 5 Approaches to Write Better Prompts
Team statworx
26. October 2022
Read more
  • Recap
  • statworx
statworx @ vuejs.de Conf 2022
Jakob Gepp
14. October 2022
Read more
  • Data Engineering
  • Data Science
Application and Infrastructure Monitoring and Logging: metrics and (event) logs
Team statworx
29. September 2022
Read more
  • Coding
  • Data Science
  • Machine Learning
Zero-Shot Text Classification
Fabian Müller
29. September 2022
Read more
  • Cloud Technology
  • Data Engineering
  • Data Science
How to Get Your Data Science Project Ready for the Cloud
Alexander Broska
14. September 2022
Read more
  • Artificial Intelligence
  • Human-centered AI
  • Machine Learning
Gender Repre­sentation in AI – Part 1: Utilizing StyleGAN to Explore Gender Directions in Face Image Editing
Isabel Hermes
18. August 2022
Read more
  • Artificial Intelligence
  • Human-centered AI
statworx AI Principles: Why We Started Developing Our Own AI Guidelines
Team statworx
04. August 2022
Read more
  • Data Engineering
  • Data Science
  • Python
How to Scan Your Code and Dependencies in Python
Thomas Alcock
21. July 2022
Read more
  • Data Engineering
  • Data Science
  • Machine Learning
Data-Centric AI: From Model-First to Data-First AI Processes
Team statworx
13. July 2022
Read more
  • Artificial Intelligence
  • Deep Learning
  • Human-centered AI
  • Machine Learning
DALL-E 2: Why Discrimination in AI Development Cannot Be Ignored
Team statworx
28. June 2022
Read more
  • R
The helfRlein package – A collection of useful functions
Jakob Gepp
23. June 2022
Read more
  • Recap
  • statworx
Unfold 2022 in Bern – by Cleverclip
Team statworx
11. May 2022
Read more
  • Artificial Intelligence
  • Data Science
  • Human-centered AI
  • Machine Learning
Break the Bias in AI
Team statworx
08. March 2022
Read more
  • Artificial Intelligence
  • Cloud Technology
  • Data Science
  • Sustainable AI
How to Reduce the AI Carbon Footprint as a Data Scientist
Team statworx
02. February 2022
Read more
  • Recap
  • statworx
2022 and the rise of statworx next
Sebastian Heinz
06. January 2022
Read more
  • Recap
  • statworx
5 highlights from the Zurich Digital Festival 2021
Team statworx
25. November 2021
Read more
  • Data Science
  • Human-centered AI
  • Machine Learning
  • Strategy
Why Data Science and AI Initiatives Fail – A Reflection on Non-Technical Factors
Team statworx
22. September 2021
Read more
  • Artificial Intelligence
  • Data Science
  • Human-centered AI
  • Machine Learning
  • statworx
Column: Human and machine side by side
Sebastian Heinz
03. September 2021
Read more
  • Coding
  • Data Science
  • Python
How to Automatically Create Project Graphs With Call Graph
Team statworx
25. August 2021
Read more
  • Coding
  • Python
  • Tutorial
statworx Cheatsheets – Python Basics Cheatsheet for Data Science
Team statworx
13. August 2021
Read more
  • Data Science
  • statworx
  • Strategy
STATWORX meets DHBW – Data Science Real-World Use Cases
Team statworx
04. August 2021
Read more
  • Data Engineering
  • Data Science
  • Machine Learning
Deploy and Scale Machine Learning Models with Kubernetes
Team statworx
29. July 2021
Read more
  • Cloud Technology
  • Data Engineering
  • Machine Learning
3 Scenarios for Deploying Machine Learning Workflows Using MLflow
Team statworx
30. June 2021
Read more
  • Artificial Intelligence
  • Deep Learning
  • Machine Learning
Car Model Classification III: Explainability of Deep Learning Models With Grad-CAM
Team statworx
19. May 2021
Read more
  • Artificial Intelligence
  • Coding
  • Deep Learning
Car Model Classification II: Deploying TensorFlow Models in Docker Using TensorFlow Serving
No items found.
12. May 2021
Read more
  • Coding
  • Deep Learning
Car Model Classification I: Transfer Learning with ResNet
Team statworx
05. May 2021
Read more
  • Artificial Intelligence
  • Deep Learning
  • Machine Learning
Car Model Classification IV: Integrating Deep Learning Models With Dash
Dominique Lade
05. May 2021
Read more
  • AI Act
Potential Not Yet Fully Tapped – A Commentary on the EU’s Proposed AI Regulation
Team statworx
28. April 2021
Read more
  • Artificial Intelligence
  • Deep Learning
  • statworx
Creaition – revolutionizing the design process with machine learning
Team statworx
31. March 2021
Read more
  • Artificial Intelligence
  • Data Science
  • Machine Learning
5 Types of Machine Learning Algorithms With Use Cases
Team statworx
24. March 2021
Read more
  • Recaps
  • statworx
2020 – A Year in Review for Me and GPT-3
Sebastian Heinz
23. Dezember 2020
Read more
  • Artificial Intelligence
  • Deep Learning
  • Machine Learning
5 Practical Examples of NLP Use Cases
Team statworx
12. November 2020
Read more
  • Data Science
  • Deep Learning
The 5 Most Important Use Cases for Computer Vision
Team statworx
11. November 2020
Read more
  • Data Science
  • Deep Learning
New Trends in Natural Language Processing – How NLP Becomes Suitable for the Mass-Market
Dominique Lade
29. October 2020
Read more
  • Data Engineering
5 Technologies That Every Data Engineer Should Know
Team statworx
22. October 2020
Read more
  • Artificial Intelligence
  • Data Science
  • Machine Learning

Generative Adversarial Networks: How Data Can Be Generated With Neural Networks
Team statworx
10. October 2020
Read more
  • Coding
  • Data Science
  • Deep Learning
Fine-tuning Tesseract OCR for German Invoices
Team statworx
08. October 2020
Read more
  • Artificial Intelligence
  • Machine Learning
Whitepaper: A Maturity Model for Artificial Intelligence
Team statworx
06. October 2020
Read more
  • Data Engineering
  • Data Science
  • Machine Learning
How to Provide Machine Learning Models With the Help Of Docker Containers
Thomas Alcock
01. October 2020
Read more
  • Recap
  • statworx
STATWORX 2.0 – Opening of the New Headquarters in Frankfurt
Julius Heinz
24. September 2020
Read more
  • Machine Learning
  • Python
  • Tutorial
How to Build a Machine Learning API with Python and Flask
Team statworx
29. July 2020
Read more
  • Data Science
  • Statistics & Methods
Model Regularization – The Bayesian Way
Thomas Alcock
15. July 2020
Read more
  • Recap
  • statworx
Off To New Adventures: STATWORX Office Soft Opening
Team statworx
14. July 2020
Read more
  • Data Engineering
  • R
  • Tutorial
How To Dockerize ShinyApps
Team statworx
15. May 2020
Read more
  • Coding
  • Python
Making Of: A Free API For COVID-19 Data
Sebastian Heinz
01. April 2020
Read more
  • Frontend
  • Python
  • Tutorial
How To Build A Dashboard In Python – Plotly Dash Step-by-Step Tutorial
Alexander Blaufuss
26. March 2020
Read more
  • Coding
  • R
Why Is It Called That Way?! – Origin and Meaning of R Package Names
Team statworx
19. March 2020
Read more
  • Data Visualization
  • R
Community Detection with Louvain and Infomap
Team statworx
04. March 2020
Read more
  • Coding
  • Data Engineering
  • Data Science
Testing REST APIs With Newman
Team statworx
26. February 2020
Read more
  • Coding
  • Frontend
  • R
Dynamic UI Elements in Shiny – Part 2
Team statworx
19. Febuary 2020
Read more
  • Coding
  • Data Visualization
  • R
Animated Plots using ggplot and gganimate
Team statworx
14. Febuary 2020
Read more
  • Machine Learning
Machine Learning Goes Causal II: Meet the Random Forest’s Causal Brother
Team statworx
05. February 2020
Read more
  • Artificial Intelligence
  • Machine Learning
  • Statistics & Methods
Machine Learning Goes Causal I: Why Causality Matters
Team statworx
29.01.2020
Read more
  • Data Engineering
  • R
  • Tutorial
How To Create REST APIs With R Plumber
Stephan Emmer
23. January 2020
Read more
  • Recaps
  • statworx
statworx 2019 – A Year in Review
Sebastian Heinz
20. Dezember 2019
Read more
  • Artificial Intelligence
  • Deep Learning
Deep Learning Overview and Getting Started
Team statworx
04. December 2019
Read more
  • Coding
  • Machine Learning
  • R
Tuning Random Forest on Time Series Data
Team statworx
21. November 2019
Read more
  • Data Science
  • R
Combining Price Elasticities and Sales Forecastings for Sales Improvement
Team statworx
06. November 2019
Read more
  • Data Engineering
  • Python
Access your Spark Cluster from Everywhere with Apache Livy
Team statworx
30. October 2019
Read more
  • Recap
  • statworx
STATWORX on Tour: Wine, Castles & Hiking!
Team statworx
18. October 2019
Read more
  • Data Science
  • R
  • Statistics & Methods
Evaluating Model Performance by Building Cross-Validation from Scratch
Team statworx
02. October 2019
Read more
  • Data Science
  • Machine Learning
  • R
Time Series Forecasting With Random Forest
Team statworx
25. September 2019
Read more
  • Coding
  • Frontend
  • R
Dynamic UI Elements in Shiny – Part 1
Team statworx
11. September 2019
Read more
  • Machine Learning
  • R
  • Statistics & Methods
What the Mape Is FALSELY Blamed For, Its TRUE Weaknesses and BETTER Alternatives!
Team statworx
16. August 2019
Read more
  • Coding
  • Python
Web Scraping 101 in Python with Requests & BeautifulSoup
Team statworx
31. July 2019
Read more
  • Coding
  • Frontend
  • R
Getting Started With Flexdashboards in R
Thomas Alcock
19. July 2019
Read more
  • Recap
  • statworx
statworx summer barbecue 2019
Team statworx
21. June 2019
Read more
  • Data Visualization
  • R
Interactive Network Visualization with R
Team statworx
12. June 2019
Read more
This is some text inside of a div block.
This is some text inside of a div block.