Jump to content

Draft:Mark I Perceptron

From Wikipedia, the free encyclopedia


The Mark I Perceptron, developed by Frank Rosenblatt in 1958, was a pioneering supervised image classification learning system. It was the first implementation of an Artificial Intelligence (AI) machine. It differs from the Perceptron which is a software architecture proposed in 1943 by by Warren McCulloch and Walter Pitts,[1] enhancements of which have continued to be an integral part of cutting edge AI technologies like the Transformer.

The Mark I Perceptron, from its operator's manual

Architecture

[edit]

The Mark I Perceptron was organized into three layers:[2]

  • A set of sensory units which receive optical input
  • A set of association units, each of which fire based on input from multiple sensory units
  • A set of response units, which fire based on input from multiple association units

The connection between sensory units and association units were random. The working of association units was very similar to the response units.[2] Different versions of the Mark I used different numbers of units in each of the layers.[3]

Capabilities

[edit]

In his 1957 proposal for development of the "Cornell Photoperceptron", Frank Rosenblatt claimed:[4]

Devices of this sort are expected ultimately to be capable of concept formation, language translation, collation of military intelligence, and the solution of problems through inductive logic.

With the first version of the Mark I Perceptron as early as 1958, Rosenblatt demonstrated a simple binary classification experiment, namely distinguishing between sheets of paper marked on the right versus those marked on the left side.[5]

One of the later experiments distinguished a square from a circle printed on paper. The shapes were perfect and their sizes fixed; only variation was in their position and orientation (rotation). The Mark I Perceptron achieved 99.8% accuracy on a test dataset with 500 neurons in a single layer. The size of the training dataset was 10,000 example images. It took 3 seconds for the training pipeline to go through a single image. Higher accuracy was observed with thick outline figures compared to solid figures, probably because outline figures reduced overfitting.[3]

Another experiment distinguished between a square and a diamond (square turned 45 degrees) for which 100% accuracy was achieved with only 60 training images (30 images each of square and diamond) with a Perceptron having 1000 neurons in a single layer. The time taken to process each training input for this larger perceptron was 15 seconds. The only variation was in position of the image, since rotation would have been ambiguous.

The same machine with 1000 neurons could distinguish between the letters X and E with 100% accuracy when trained with only 20 images (10 images of each letter). Variations in the images included both position and rotation by up to 30 degrees. When variation in rotation was increased to any angle (both in training and test datasets), the accuracy reduced to 90% with 60 training images (30 images of each letter).[3]

For distinguishing between the letters E and F, a more challenging problem due to their similarity, the same 1000 neuron perceptron achieved an accuracy of more than 80% with 60 training images. Variation was only in the position of the image, with no rotation.[3]

References

[edit]
  1. ^ McCulloch, Warren S.; Pitts, Walter (1943-12-01). "A logical calculus of the ideas immanent in nervous activity". The Bulletin of Mathematical Biophysics. 5 (4): 115–133. doi:10.1007/BF02478259. ISSN 1522-9602.
  2. ^ a b Rosenblatt, F. (1958). "The perceptron: A probabilistic model for information storage and organization in the brain". Psychological Review. 65 (6): 386–408. doi:10.1037/h0042519. ISSN 1939-1471. PMID 13602029.
  3. ^ a b c d Rosenblatt, Frank (March 1960). "Perceptron Simulation Experiments". Proceedings of the IRE. 48 (3): 301–309. doi:10.1109/JRPROC.1960.287598. ISSN 0096-8390.
  4. ^ Rosenblatt, Frank (January 1957). "The Perceptron—a perceiving and recognizing automaton" (PDF). Cornell Aeronautical Laboratory.
  5. ^ "Professor's perceptron paved the way for AI – 60 years too soon | Cornell Chronicle". news.cornell.edu. Retrieved 2024-10-08.