Search This Blog

AI - The Beginning

AI’s history is longer and more uneven than is often assumed.

The field traces its intellectual roots to Alan Turing’s 1950 essay “Can Machines Think?”. In 1951, researchers ran the first working AI program on the Ferranti Mark 1 at the University of Manchester. In 1956, a team at Carnegie Mellon University in Pittsburgh produced the “Logical Theorist”, the first program built to perform automated reasoning.

Attention then shifted toward expert systems. Development began in 1964 and culminated in 1969 with DENDRAL, which assisted organic chemists in designing complex syntheses. Its performance depended on rules extracted from human laboratory experience. In 1971, MYCIN extended this approach to medical diagnosis and treatment using the LISP language. It worked best in the hands of skilled doctors who could resolve uncertainty by ordering further tests.

These successes prompted widespread adoption of expert systems throughout the 1970s and early 1980s. Most followed a common architecture based on a knowledge base and an inference engine. By the mid-1980s, this approach had reached its limits. Capturing expert knowledge in rule form proved difficult, and systems became increasingly complex. The interaction of forward chaining and backward chaining made them powerful but fragile, contributing to the decline of the first expert system era.

Expert systems were ultimately seen as slow and expensive to develop. Even so, they remain valuable in settings where transparency and review are required, such as credit scoring and job application screening, and in situations where fast execution is essential, including self-driving vehicle systems.

In parallel, two other experimental approaches were pursued, often by researchers working outside engineering. In 1957, Frank Rosenblatt, a psychologist, introduced the “Perceptron”, an early attempt to build a system capable of learning by trial and error. This connectionist model represented the first neural network. It showed limited success in image recognition, including distinguishing cats from dogs, but failed to secure continued funding. Another early initiative focused on machine translation, yet poor performance led to its termination in 1966.

These disappointments, combined with poorly structured data, hardware constraints, and limited processing power, led to a loss of confidence and funding. This period of decline, from 1974 to 1980, became known as the First AI Winter.

A cautious revival followed with explanation-based learning. This approach depended on a human expert who explained how a specific case could support a general rule. It enabled faster systems, especially when data was sparse or unreliable, but remained constrained by the quality of human input. Its strengths were focus, efficiency, and traceability, which made it suitable for use in legal reasoning systems.

In 1983, further progress was made with the development of an effective recurrent neural network. Repeated activation between connected units increased sensitivity, allowing patterns of activity to stabilise in a manner comparable to learning.

This improvement to the expert system concept resulted in earlier systems becoming obsolete and 1987 saw the final collapse of the earlier LISP-based expert systems and the start of the “Second AI Winter” (1987 – 2000).

Development did of course move along during this so-called “winter”. Of note were Support Vector Machines designed to improve data classification and regression analysis, of which the dog/cat sorting is a trivial example. In 1995 these were enhanced by the introduction of “Random Forests” – ensembles of decision trees based on random subsets of the presented data from which a corresponding number of predictions were made, which were then averaged or a majority vote taken on the classification to get the final result. Shortly after that, in 1996, the oddly named Long Short Term Memory (LSTM) units were incorporated into recurrent neural networks which, if unmodified, had a tendency to lose data that might be helpful later on in the learning process. The LSTMs took decisions on retaining or forgetting earlier assessment of data and if retained, doing so over hundreds of time intervals. So some data was indeed discarded but other was retained over a longer timescale that was represented in conventional “short term memory”. The practical effect was greater sensitivity to processing results which the LSTMs judged likely to be useful later on, at the expense of other results judged expendable. However this was expressed in terms of improved speed and accuracy.

During the next 13 years development concentrated on market testing robotic products like Roomba, a domestic robot vacuum cleaner and robot grass cutters while image processing was improved to sort and standardize machine readable images from large databases and for use in self-driving vehicles.

2009 saw the first LSTM recurrent neural network with pattern recognition software and this enabled cursive hand writing to be read and Google built an “autonomous car”. At the same time the ImageNet visual dataset containing 14 million hand annotated images was produced using a team of 49,000 individuals from 167 countries working from a base of 162 million candidate images!

The decade that followed brought fast-moving innovation. In 2013, Google significantly improved natural language processing, helping to establish the chatbot as a practical tool. Generative AI then expanded machine output beyond prediction to prompt-led content creation, producing fluent and often convincing text. Meanwhile, the now familiar AI-generated images from OpenAI’s DALL-E 3 became part of everyday digital culture.

In 2017, OpenAI released influential research on generative models. At the same time, Google DeepMind in London refined AlphaGo into AlphaGo Zero, which learned to play Go entirely through self-play. The same approach was applied to chess, with similarly strong results against both machines and human opponents.

In 2018, Google Duplex showed that an AI assistant could make telephone bookings in real-world settings. The following year saw the launch of OpenAI’s GPT-4, which was widely praised despite continued concerns about fabricated responses inherited from earlier models.

ChatGPT was released in November 2022 and quickly became a public reference point for AI. Its ongoing hallucination problems sparked political debate. In parallel, a wave of legal challenges emerged against newer AI companies, often centred on copyright infringement and the unauthorised use of private or personal data in training sets.

Late in 2024, President Joe Biden issued an executive order defining eight goals for ethical AI development in the United States. These included protecting national interests, respecting copyright, safeguarding personal data, and ensuring AI systems were accurate and non-discriminatory.

This order was rescinded by Donald Trump on his first day in office, removing regulatory obligations for US AI firms. This is likely to deepen legal conflict between major companies while discouraging public challenge.

On 10 and 11 February 2025, France hosted the Artificial Intelligence Action Summit. Sixty-one countries signed a declaration supporting inclusive and sustainable AI. The UK and the US declined. Anglo-Saxon exceptionalism?




Find More Articles on Bytes & Pieces


How We're Redesigning Audacity For The Future

Audacity is being upgraded by the folks who run Muse. This video digs into the research and design that went into gradually transforming Audacity 3 - a free and open source audio editor and production app - and how they've built on that work to develop Audacity 4. 

After some 3,500 comments, I asked Gemini to review a text file containing all those comments and spit out a review. The review is beneath the video, which, incidentally is very well made, easy to follow, and in my view an excellent no-messing about lesson on how to do project development, especially on a product which has been in the public domain for decades.


 

Overall Sentiment


The comments reflect a strong dichotomy: widespread criticism of the new logo and branding, but overwhelming praise for the planned software improvements, the video's transparency, and the development team's approach. Most commenters are excited about the future of the software itself, even if they dislike its new visual identity.

---

 1. Intense Feedback on the New Logo and Rebranding

current branding






The most discussed topic by a significant margin is the rebranding, with most of the feedback being critical.

new suggestion in the video








* Loss of the Waveform: A vast number of users feel the waveform was the most iconic and essential part of the original logo, even more so than the headphones. Many believe its removal strips the logo of its identity and fails to communicate what the program does. Suggestions were frequently made to reincorporate a simplified waveform.
* Color Change: The shift from blue to red was jarring for many commenters, who strongly associate the color blue with Audacity's brand identity. They argue that the new red color feels alien and corporate.
* Aesthetic and Design: The new logo is frequently described as "lifeless," "sterile," "corporate," and "soulless". Some find the asymmetrical design to be visually unbalanced, awkward, and uncomfortable to look at. Others feel it looks like a generic stock icon that lacks personality.
* Loss of Recognisability: Many feel the new logo is so different that it's unrecognisable as Audacity, abandoning the charm and identity of the original. Several commenters compared it to other controversial corporate rebrands.
* Contextual Understanding: A smaller group of commenters noted that while the logo is strange on its own, it makes more sense and fits in well when viewed alongside the other apps in the Muse suite. Some simply expressed that they like the new modern look.

---

 2. High Praise for the Video and the Presenter's Approach


There is near-universal acclaim for the video itself and the way the project is being managed and communicated.

* Transparency and Honesty: Commenters repeatedly praised the video's transparency, depth, and honesty about the development process, including the challenges of technical debt. Many wish other software projects, especially open-source ones, had this level of communication.
* Excellent Communication: The presenter is described as an inspiring and skilled leader who is excellent at communicating complex design principles and the reasoning behind decisions.
* Addressing Criticism: Viewers appreciated the direct response to poorly-researched, negative videos that criticized an early, unfinished build of the software.

---

 3. Excitement and Support for Audacity 4.0's New Features


The actual changes to the software detailed in the video were met with widespread excitement and approval.

* UI/UX Improvements: Users are very enthusiastic about the modernized UI and thoughtful UX improvements. Many expressed that the video addressed long-standing frustrations they've had with the software for years.
* Specific Features: There was particular excitement for the introduction of non-destructive effects, the new "Split" tool, and the plan to eliminate confusing "modes". The solution of asking the user to choose their preferred delete behavior was lauded as a brilliant UX decision.
* Technical Decisions: The development team was commended for choosing to refactor the existing codebase rather than starting over and for migrating to Qt instead of a web-based framework like Electron.

---

 4. Diverse and Creative Use Cases Mentioned by Users


The comments revealed the vast and sometimes unexpected ways people use Audacity.

* Standard Uses: Common uses mentioned include recording voice-overs, simple audio editing, cleaning up recordings, academic analysis, and as a first step into music production.
* Unconventional Uses: Several users mentioned creative and unusual applications, such as datamoshing and glitch art by importing raw image data (.bmp, .jpeg) as audio, applying effects, and exporting it back as an image.
* Niche Applications: Other uses include a lawyer using it to prepare audio evidence for court, creating maps for a rhythm game, and playing ARGs.

---

Audacity: Link


Find More Articles on Bytes & Pieces


Dynamic Pattern Generator




Find More Articles on Bytes & Pieces


Using Google Drive Images in Scripts

Option 1: Direct Download Link (Simplest)

Right-click the image in Google Drive and select "Share"
Set sharing to "Anyone with the link" (if appropriate for your use case)
Get the share link

Modify the link to create a direct download URL:
Original format: https://drive.google.com/file/d/FILE_ID/view?usp=sharing
Download format: https://drive.google.com/uc?export=download&id=FILE_ID

In your script, you can then use this URL directly:

# Python example
import requests

image_url = "https://drive.google.com/uc?export=download&id=YOUR_FILE_ID"
response = requests.get(image_url)

if response.status_code == 200:
    with open('downloaded_image.jpg', 'wb') as f:
        f.write(response.content)


So,
https://drive.google.com/file/d/1Y0Y7WDpmRH0YAH6dcMG6Cul8jA5WJus2/view?usp=drive_link
becomes
https://drive.google.com/uc?export=download&id=1Y0Y7WDpmRH0YAH6dcMG6Cul8jA5WJus2

Option 2: Google Drive API (More Robust)

For more control and security, use the Google Drive API:

Set up a Google Cloud Platform project
Enable the Google Drive API
Create credentials (OAuth client ID or service account)
Install the Google Client Library for your language

# Python example using Google Drive API
from googleapiclient.discovery import build
from google.oauth2 import service_account

# Authenticate
credentials = service_account.Credentials.from_service_account_file(
    'service-account.json', scopes=['https://www.googleapis.com/auth/drive.readonly'])
drive_service = build('drive', 'v3', credentials=credentials)

# Get the file
file_id = 'YOUR_FILE_ID'
request = drive_service.files().get_media(fileId=file_id)
response = request.execute()

# Save the file
with open('downloaded_image.jpg', 'wb') as f:
    f.write(response)

Option 3: Google Drive Desktop Sync

If your script runs on a computer with Google Drive desktop sync:

Ensure the image is synced to your local Google Drive folder
Reference the file using its local path in your script

# Python example
import shutil

local_path = "/path/to/your/google/drive/image.jpg"
project_path = "/path/to/project/default_image.jpg"

shutil.copy2(local_path, project_path)

Find More Articles on Bytes & Pieces


Upload Your Picture for Kaliedoscope - Google Gemini 2.5PRO

Interactive Kaleidoscope

Kaleidoscope

Create mesmerizing patterns from your images.

Counter-Clockwise Still Clockwise
Zoom Out Zoom In

Created with Google Gemini 2.5PRO at https://gemini.google.com/ using the instruction: " Create a "kaleidoscope" webpage. Users will be able to upload their own picture which you will edit to be circular and placed in the center of a scrollable web page limited to a maximum of 1260px wide. If users do not have a picture to upload, they will be able to select the picture Included with these instructions. Most striking images occur when the original picture is turned slightly. The user will therefore need to be able to make subtle changes to the speed and direction of rotation of their image. The user will be able to zoom the image."


Find More Articles on Bytes & Pieces


Upload your picture for a Kaleidoscope - Z.AI GLM4.5

Interactive Kaleidoscope

Interactive Kaleidoscope

Upload your own image or use our default image to create a mesmerizing kaleidoscope effect. Adjust the speed, zoom, and direction to create the perfect visual experience.

2.0
1.0

Tip: Combine different zoom levels with rotation speeds for unique kaleidoscope patterns!

Created with GLMP4.5 at https://z.ai/ using the instruction: " Create a "kaleidoscope" webpage. Users will be able to upload their own picture which you will edit to be circular and placed in the center of a scrollable web page limited to a maximum of 1260px wide. If users do not have a picture to upload, they will be able to select the picture Included with these instructions. Most striking images occur when the original picture is turned slightly. The user will therefore need to be able to make subtle changes to the speed and direction of rotation of their image. The user will be able to zoom the image."


Find More Articles on Bytes & Pieces


Bubble Pop - Made Today with GLM 4.5

GLM 4.5 is at z.ai Digital Bubble Wrap - Pop Away Your Stress!

🫧 Digital Bubble Wrap

Click the bubbles to pop them and relieve your stress!

0 Bubbles Popped
0 Current Combo
0 Best Combo
>


Find More Articles on Bytes & Pieces


Bubble Pop - Made today with Google Gemini

Made today with Google Gemini Digital Bubble Wrap

Digital Bubble Wrap

Click the bubbles to pop them. Enjoy the satisfaction!




Find More Articles on Bytes & Pieces


Using ChatGPT to Help Sort Out A Divorce

You:
I live in the UK. Married for 10 years and now my wife wants a divorce. I bought the house we live in 30 years ago. Before we married my outstanding mortgage was £14000. My wife had a home of her own which she sold before marriage but it left her with negative equity of £20000. To cover that, we agreed to convert my mortgage to a joint mortgage of £34000. We have two children we both love dearly. What would be the best way forward if we divorce?

Interactive Drum Machine

16-Step Drum Machine

16-Step Drum Machine

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
Hi-hat
Snare
Kick
120 BPM


Find More Articles on Bytes & Pieces


Collide Colours

Large Color Collider

Large Color Collider




Find More Articles on Bytes & Pieces