For the most part, when you have a web site, you want search engines to find it, but there may be parts you don’t want them to crawl because there are duplicate pages, or scripts that will put a heavy load on the database. There are many legitimate reasons you might want to do this.
That’s where a robots.txt file comes in. There’s an entire specification for robots.txt, documenting what rules the various responsible search spiders will follow.
The problem is, robots.txt is voluntary. Not all the search search spiders will obey it. Right now for example, there’s a whole world of AI crawlers trying to suck down the contents of the entire Internet.
Sometime in the mid-to-late 80s, my brothers and I got hooked on a very early text adventure game called “Camel”. The premise is that you’re fleeing across the desert on a camel, being chased by “pygmies.”
I don’t recall how we obtained the game; I do remember loading it from a floppy disk, but I’m not sure whether we bought the disk preloaded or if someone typed it in from a magazine, but we played it enough times that some of the program’s outputs have been a running joke in my family for decades.
The line “Your camel thanks you” (displayed when the player chose to stop for the night) came up again this weekend. This inspired me to see if I could find the BASIC program online.
I’m putting together a small collection of pages to share with family. A CMS would be overkill for this, so I’m creating the HTML by hand (and sometimes I enjoy doing this stuff by hand).
I’ve reached the point where I’m adding text, describing the photos and the memories they bring back. And because I’m using static HTML files, I’m using Visual Studio Code.
The problem is, the GitHub Co-Pilot that’s useful for auto-completing some of the boilerplate, also wants to chime in and offer suggestions for the text of my memories. Not only is this distracting, but some of the text is wildly inaccurate (after all, these are my memories, not the bot’s). So, how do you turn this off?
It turns out this is relatively simple:
Go into the Command Palette (Ctrl-Shift-P or Command-Shift-P)
Did you know the MacOS “Preview” tool can be used to capture the text in a photo of a sign? All you do is click on the first word, and then drag the mouse until all the text is selected. Press Command-C, and then you paste the text into another document. (The Windows Snipping tool has similar functionality, though I’ve not tried it recently.)
I’ve been working on a small photo site to share with my family. One of the photos is of the sign in front of a garden for the blind and I want to make the text easily readable outside the photo.
The thing with this photo is, it has a lot of text. And I mean, a LOT. (Go ahead and click the image below, you’ll see what I mean.)
I’m really impressed with the preview tool. Two click and drag operations and I had 95% of the text. All I had to was type the context of. one paragraph, and then do a bit of proofreading (plus remove some hyphens where the text no longer had line breaks.)
Less than an hour of work.
Curious what the sign says? Here’s the text
Welcome to Wynnewood Valley Parks’s Sensory Garden
The Garden’s Origins…
The Rotary Club of Ardmore, in partnership with the Township of Lower Merion; built and dedicated the Wynnewood Valley Park. Garden for the Blind and Children” in 1977.
Ardmore Rotarian and professional landscaper, the late Bill Mulford, spearheaded planning and construction of the multi-tièred garden which introduced Braille plates to identify the individual flowers and shrubs. Prominent among the plantings were 25 cherry trees, a gift from the Rotary Club of Tokyo, Japan in appreciation for a 1973 gift of 300 dogwood saplings from the Ardmore Rotary Club for the Imperial Gardens. The saplings replaced many of the.garden’s 50 dogwood trees which had been damaged during WW II. The 50 dogwood trees, a 1912 gift from the United States of America, were in appreciation for 3000 cherry trees from. Japan which were planted earlier along the Potomac River in Washington Đ.C.
Ardmore Rotarians and the Men’s Garden Club of Delaware Valley maintained the Garden for the Blind until the Garden Club members were unable to continue, In 2005, the Rotary Club of Ardmore, under the leadership of Past President Dr. Douglas Klepfer, decided to restore the Garden, committed funds and began to build a new partnership with Lower Merion Township, Lower Merion Conservancy, the Penn Wynne community and others. A major contribution was presented by The Kistler-Tiffany Foundation to purchase all the plants.
The result is a beautiful new sensory garden and park entrance walkway accessible for all to enjoy. Included in the landscaping are the 12 new “Cherry Trees of Goodwill” donated by the Rotary Club of Tokyo in appreciation for Ardmore Rotary’s gift of 200 pink and white dogwood trees to the Rotary Club of Tokyo in 2004. The trees have been planted in the Imperial Gardens and throughout Japan. The Lower Merion Conservancy has committed its services for garden education and maintenance.
The restored Garden exemplifies the “spirit of the flower that binds the world together” as the Tokyo Rotary Club concluded when accepting the 2004 gift of dogwood trees. Planted in late 2010, the garden celebrated its ribbon-cutting ceremony on May 9, 2011.
The Plants in the Garden…
The Sensory Garden bas 73 species of plants including herbaceous plants and trees and shrubs.
There are some scientific and medical scenarios (e.g. particle detection) that require instruments made with steel with as little radioactive contamination as possible. This is challenging as atomic-bomb testing during the 20th century resulted in persistent airborne nuclear fallout, and so there’s apparently some industry in salvaging steel from ships which sank prior to the Trinity tests. This is known as “Low-background steel.“
“AI Slop” is an increasingly popular term to describe the growing amount of AI-generated content on the web, much of it intended to draw page views, spread falsehoods, and the like.
Last week, I learned of the “Low-background Steel” blog which is dedicated to sources of data from before the end of 2022 (because ChatGPT premiered in November of that year). At the moment, it only has a half-dozen entries.
I recently encountered a list of “Falsehoods programmers believe about time.” With the possible exception of the one about black holes (third from the end), they’re all false.
“There are always 24 hours in a day”? Not when we “spring forward” to Daylight Saving Time; that day only has 23 hours. Likewise, when we “fall back” to Standard Time, that day has 25 hours. Of course, even those rules aren’t true if you live in a region that doesn’t observe Daylight Saving Time.
Even Google has trouble with this. Monday morning, at 9:07 AM Eastern Daylight Time, I asked Google what time zone Colorado is in. The response was Mountain Time, which is what I was expecting.
Google then offered a little extra information. In Maryland, it was 9:07 AM Monday morning. But in Colorado, according to Google, it was 8:46 AM on Sunday.
That time zone offset is a bit more complicated than I expected!
As part of the book tour for his new Book, Class Clown: The Memoirs of a Professional Wiseass, Dave Barry sat for an interview at a Microsoft in-house event. The interviewer was Scott Hanselman, who is also Microsoft’s Vice President of Developer Community (so, a “tech-influencer”).
It’s one of the funniest interviews I’ve heard recently. The spleen story alone is enough to make it worth your time. And now you’ll have to play it too, if for no other reason than to find out about this “spleen story.”
Steve was a little confused by this mention of entering a two-digit code, but it was something I immediately recognized, as did co-host Leo Laporte.
Here’s a clip from this week’s show (episode 1026).
If you didn’t click the video, you’re depriving yourself of the chance to hear Steve Gibson reading an email I sent him.
Hi Steve,
I just listened to episode 1025 in which you read a bit of listener feedback that left you perplexed about Microsoft’s Authenticator app needing you to type in a two-digit number. I use Microsoft’s products in an enterprise environment and thought I might be able to shed some light on this.
What’s going on is that Microsoft offers the option of using a push notification instead of the TOTP (in the enterprises I’m familiar with allow you to use either as a second factor).
The problem with push notifications is, of course, “notification fatigue.” People get used to seeing the notification and just click “Yes, it’s me” without thinking it through. (So if someone figures out your password, and you blindly confirm it…. I’m sure you see where that’s going.)
To counter this, when you log in to a Microsoft system that uses push notifications, they display a two digit number. You then have to enter that number into the pop-up from the authenticator app. That way, it’s much more difficult for an end-user to accidentally confirm a third-party’s login attempt.
I tend to take a lot of somewhat random photos of things that interest me. If it’s a good photo and the mood strikes, I’ll share it on various social media channels.
One image I recently shared featured a stack of “Wooden Bitcoin” with a joke about their value (there are certain scenarios where the wooden ones are quite valuable).
When I shared the image on social media (and when I embedded it above), I included an image description for accessibility purposes.
A small pile of wooden disks on a silver background. An outer circle on each disk reads “Wooden Bitcoin. Just as Good, Really.” Text on an inner circle reads “To Redeem: Insert into any Internet Coin Slot.” The Bitcoin logo appears at the center of each disk.
That description wasn’t very hard to come up with, but sometimes, it’s a bit more difficult to convey all the visual details. For example, consider this image of winterberries.
In the embedded image’s alt text I’ve described it as “Bright red winterberries on bare branches” which is fairly accurate, though it doesn’t offer a lot of detail.
Much like everyone else, I’ve been getting my feet wet with AI. Lately, I’ve been using Ollama to run models locally for various toy projects; and more recently, I noticed that Meta’s llama3.2-vision model had become available, allowing the ability to work with image files.
Uploading the winterberry image, I gave llama the prompt:
Please describe this image.
What I got back was quite a bit more detailed than what I’d come up with on my own.
This image presents a close-up view of a vibrant red berry bush, showcasing its intricate details.
The bush’s branches are bare and grayish-brown in color, with numerous small twigs extending from the main stems. The berries themselves are bright red, spherical in shape, and densely packed along the branches. They appear to be ripe, as they have lost their green hue and turned a deep crimson color.
In the background, other bushes or plants can be seen, although they are not clearly visible due to the dense foliage of the berry bush. The overall atmosphere suggests that this image was captured during the winter season, likely in December or January, given the bare branches and ripe berries.
That description pretty well knocks it out of the park! And although the image contains no Exif data, the AI correctly determined that the photo was taken “…in December or January” (December 31 in fact, right on the cusp of the change).
For comparison, I uploaded the same image to ChatGPT, using the current “GPT-4o mini” model. Entering the same prompt, I got a similarly detailed, albeit shorter, description.
This image shows branches covered with clusters of bright red berries. The berries are small, smooth, and shiny, growing densely along the thin, leafless twigs. These are likely the fruits of a holly plant, possibly winterberry (Ilex verticillata), which is known for its vibrant red berries that remain on the branches during winter, adding color to the landscape. The background is blurred, emphasizing the striking red berries in the foreground.
Again, that’s a great description. No guesses on when the photo was taken, but this time, ChatGPT correctly recognized that these were winterberries.
In the case of llama 3.2, I’d likely remove the guess about when the photo was taken since I know that information. For ChatGPT, I’d take out the guess work about the type of berries since I already know it. But to be honest, either of these outputs would be good starting points for describing the photo.
So, what about that first photo with the wooden bitcoins?
Here’s the llama output. It’s wordy and repeats some the details multiple times. And I disagree with its description of the Bitcoin logo as the curved lines form the letter “B” and not a circle.
The image shows a collection of wooden coins with the Bitcoin logo and text on them, likely used as tokens or rewards in an online game or platform.
The coins are made of wood and have a circular shape.
They feature the Bitcoin logo in the center, which is a stylized letter “B” made up of two curved lines that form a circle.
The text on the coins reads “WOODEN BITCOIN” at the top and “JUST AS GOOD, REALLY” at the bottom.
There are 15 coins visible in the image, arranged randomly on a gray surface.
Some of the coins overlap each other, creating a sense of depth and layering.
The background is a plain gray color, which helps to focus attention on the coins.
Overall, the image suggests that these wooden coins are being used as a novelty item or collectible, perhaps in conjunction with an online game or platform that rewards players with virtual currency.
Chat GPT did better on this one.
This image shows a pile of wooden tokens designed to resemble bitcoins. Each token is circular and engraved with text and imagery related to Bitcoin. The engraving includes the Bitcoin logo (a “B” with two vertical lines) and phrases such as “Wooden Bitcoin,” “Just As Good, Really,” and “To Redeem: Insert into any internet coin slot.” The tokens are humorously crafted to mimic physical representations of digital currency.
In this case, Chat GPT clearly did the better job, and both tools did a credible job of recognizing text presented in an unusual format, but I’m not sure I’d want to trust either one to generate text without reviewing it before sending the text into the world.
This is a great step toward solving the “blank page problem”, most likely giving a solid starting point toward a description of the image (it’s also a huge step up from xkcd’s “Tasks” cartoon).
At the moment, my phone’s camera roll has nearly 10,000 photos on it (I really do take a lot of somewhat random photo). Being able to run them all through AI has the potential to save a lot of tedium with creating descriptions. But, along with the need to double-check the descriptions, I’m also still on the the hook for deciding whether the image is even worth sharing.
Using an AI tool definitely has potential to save me some time. But it’s not a panacea.
Initially, you could just go to https://www.linkedin.com/mypreferences/d/settings/data-for-ai-improvement to turn off the setting, but clicking that link doesn’t seem to be reliable. (If I go directly to that link, it works, but if it goes through Facebook’s link tracker, it goes to a page not found error.)
In a desktop web browser, you can click on “View Settings” in the left navigation, click on “Data Privacy” and then, under the “How LinkedIn uses your data” heading, click on “Data for Generative AI Improvement.” At the moment, that page has a single toggle for “Use my data for training content creation AI models.”
I don’t have the app installed myself, but I’m told you can similarly go to Settings -> Data Privacy -> How LinkedIn Uses Your Data -> Data for Generative AI Improvement