ML on Apple ][+

mdcramer.github.io

119 points by mcramer 2 days ago


shagie - 2 days ago

One of my early "this is neat" programs was a genetic algorithm in Pascal. You entered a bunch of digits and it "evolved" the same sequence of digits. It started out with 10 random numbers. Their fitness (lower was better) was the sum the difference. So if the target was "123456" and the test number was "214365", it had a fitness of 6. It took the top 5, and then mutated a random digit by a random +/- 1. It printed out each row with the full population. and so you could see it scrolling as it converged on the target number.

Looking back, I want to say it was probably the July, 1992 issue of Scientific American that inspired me to write that ( https://www.geos.ed.ac.uk/~mscgis/12-13/s1100074/Holland.pdf ) . And as that was '92, this might have been on a Mac rather than an Apple ][+... it was certainly in Pascal (my first class in C was in August '92) and I had access to both at the time (I don't think it was turbo pascal on a PC as this was a summer thing and I didn't have a IBM PC at home at the time). Alas, I remember more about the specifics of the program than I do about what desk I was sitting at.

aperrien - 2 days ago

An Aeon ago in 1984, I wrote a perceptron on the Apple II. It was amazingly slow (20 minutes to complete a recognition pass), but what most impressed me at the time was that it did work. Since that time as a kid I always wondered just how far linear optimization techniques could take us. If I could just tell myself then what I know now...

windsignaling - 2 days ago

I'm surprised no one else has commented that a few of the conceptual comments in this article are a bit odd or just wrong.

> The final accuracy is 90% because 1 of the 10 observations is on the incorrect side of the decision boundary.

Who is using K-means for classification? If you have labels, then a supervised algorithm seems like a more appropriate choice.

> K-means clustering is a recursive algorithm

It is?

> If we know that the distributions are Gaussian, which is very frequently the case in machine learning

It is?

> we can employ a more powerful algorithm: Expectation Maximization (EM)

K-means is already an instance of the EM algorithm.

aardvark179 - 2 days ago

I thought this was going to be about the programming language, and I was wondering how they managed to implement it on a machine that small.

JSR_FDED - 2 days ago

Applesoft BASIC is just so darn readable. Youngsters have nothing comparable these days to learn the basics of expressing an algorithm without having to know a lot more.

And if it ever became too slow, you could reimplement the slow part in 6502 assembler, which has its own elegance. Great way to learn, glad I came up that way.

drob518 - 2 days ago

Upvoted purely for nostalgia.

amilios - 2 days ago

Bit of a weird choice to draw a decision boundary for a clustering algorithm...

alexshendi - 2 days ago

This motivates me to try this on my Ministrel 4th (21th century Jupiter Ace clone).

- 2 days ago
[deleted]
nikolay - a day ago

You don't even need a computer for ML [0]!

[0]: https://proceedings.mlr.press/v170/marx22a/marx22a.pdf

gwbas1c - 2 days ago

Any particular reason why the author chose to do this on an Apple ][?

(I mean, the pictures look cool and all.)

IE, did the author want to experiment with older forms of basic; or were they trying to learn more about old computers?

hyperliner - 2 days ago

[dead]

rob_c - 2 days ago

Since when did regression get upgraded to full blown ML?