little ai can’t write python code

Two posts back I wrote about installing GPT4All on my M1 MacBook Pro. At first blush I was impressed, as it seemed to answer the softball questions I asked it correctly and with style. Then I got a little more serious and asked it this question:

write a python program that reads a directoryand prints out everything it finds

To which GPT4All Falcon gave this answer:

Here's an example Python program that uses the`os` module to read a directory and print outeverything it finds:
import os# Set the directory pathdir_path = '/path/to/directory'# Check if the directory existsif not os.path.exists(dir_path):print("Directory does not exist")return# Iterate through all files in the directoryfor file in os.listdir(dir_path):# Check if the file is a directoryif os.path.isdir(os.path.join(dir_path, file)):# If it's a directory, recursively call this functionprint("Directory:", file)print("Files:", os.listdir(os.path.join(dir_path, file)))# Check if the directory is emptyif not os.path.exists(dir_path):print("Directory does not exist")
This program uses the `os.listdir` functionto get a list of all files in the directoryand the `os.path.join` function to constructthe full path for each file.It then calls itself recursively on anydirectories it finds, printing out thedirectory name and its contents.

Let’s examine the Python code. There is a blatant error on line 9. Python does not allow the use of a return outside of a proceedure. If you need to exit a script due to some condition, then you use Python’s exit keyword.

The second problem are lines 19-21. They don’t do anything. The message on line 21 is the same as the message on line 8. If the block of code in lines 19-21 is to determine if the directory is empty then it should say that.

Let’s clean up the code enough to actually work.

#!/usr/bin/env python3import os# Set the directory pathdir_path = '.'# Check if the directory existsif not os.path.exists(dir_path):print("Directory does not exist")exit(1)# Iterate through all files in the directoryfor file in os.listdir(dir_path):# Check if the file is a directoryif os.path.isdir(os.path.join(dir_path, file)):# If it's a directory, recursively call this functionprint("Directory:", file)print("Files:", os.listdir(os.path.join(dir_path, file)))

Now let’s test it. I’ve created some dummy directories and files. Here’s what the utility tree shows.

├── foo│   ├── bar│   │   ├── five│   │   ├── seven│   │   └── six│   ├── four│   ├── one│   ├── three│   └── two├── foo2

Now let’s run our corrected AI code and look at the output.

Directory: fooFiles: ['three', 'one', 'two', 'bar', 'four']Directory: foo2Files: []

It’s unsurprisingly disappointing. There’s a comment on line 15 (16 in the corrected code) that says to “recursively call this function” if a directory is detected. It doesn’t in part because there are no functions, and because the test for a directory doesn’t do anything but print out if it’s a directory.

The code looks like during its learning phase it swept up someone’s Stack Overflow question, but didn’t know enough to look at any answers, assuming there were any. Whatever ideas I might have had to incorporate the Falcon model into anything have been pretty much squelched. I’ll install some others and give them a spin.

Links

ChatGPT Isn’t Coming for Your Coding Job — Wired — https://www.wired.com/story/chatgpt-coding-software-crisis/

gpt4all on an m1 macbook pro

I’ve been tinkering a bit with large language models small enough to run locally on my M1 MacBook Pro without having to reach out across the internets. Local LLMs means not having to send anything across that might be collected, and thus losing control on whatever I type in (not that whatever I type in is worth collecting…).

You can go look at GPT4All here: https://gpt4all.io/index.html

The application, when first installed, can’t do anything. You need to begin to add LLMs via a curated selection it presents. These are the eleven currently listed.

  1. Wizard V1.1
  2. GPT4All Falcon (the one I’m currently using)
  3. Hermes
  4. ChatGPT-3.5 Turbo
  5. ChatGPT-4
  6. Snoozy
  7. Mini Orca
  8. Mini Orca (small)
  9. Mini Orca (large)
  10. Wizard uncensorted
  11. Llama-2-7B Chat

I’ve barely scratched the surface, having just installed the application and then selected a given model. My criterial right now for selecting a model is that there is no link back to the cloud and that it can run in the “limited” 16 GiB of memory on my MacBook. GPT4All Falcon, for example, will run in 8 GiB.

If you stop and think about the resources required by Falcon it suddenly becomes apparent that it could run on an Apple iPhone if the iPhone had enough memory. Today’s contemporary iPhones certainly have the processor horsepower and more than enough local storage to hold such a model. The problem is that the iPhone 15, Apple’s latest device, doesn’t. Perhaps the iPad Pros running with either the M1 or M2 has enough, but I know the iPhones do not.

Why would I want that much memory? To run one of the LLMs locally. And why do that? So that the LLM could “learn” about me from all the data that passes through my iPhone from me, and help the iPhone to be an even better aid. Smartphones, especially the iPhone, has long since passed the point of “good enough” to perform such tasks. The big question is how much power will that require, because there are always trade offs with portable computers. The more hardware you have on an iPhone, such as more memory, the more power the device will consume. The more processing the CPUs on the iPhone needs to perform, the more power the device will consume. As the saying goes, be careful what you ask for.

And there’s another reason for looking at this for local processing, and that’s in support of the intelligent home. My wife and I grow progressively older. We’re both in our early 70s. I would like a home with sensors and microphones and speakers to be able to assist our living here, rather than have to go live in a classical assisted living facility. And I do NOT want devices that require cloud connectivity and a monthly fee. Right now I have an intelligent [sic] doorbell which is absolute shit. I want far better, and I want local autonomy.

I haven’t even scratched the surface with any of this software. But it’s now approachable and I can run it locally on my MacBook, which I find absolutely amazing. I hope I can do more with it going forward.