Management? Never Been Easier

April 4, 2025

Age-Old Management Techniques That Still Happen

Millennials and Gen Z focus on prioritizing mental health and not dealing with the same BS as earlier generations. Some might see these priorities as infantile, but even past generations didn’t like them. INC breaks down “4 Really Bad Management Behaviors That Drive Good Employees Away” and the article really drives home how these management techniques have been in place for eons.

There article’s author Marcel Schwantes starts with a very wise statement: “Not everyone in management should be there.” He then breaks down four poor management techniques…er…behaviors that employees have been fed up with since forever. The first is shooting down their workers’ ideas. This happens on a daily basis and no one in the modern workforce won’t experience this at some point in their careers. Sometimes ideas need to be shot down, but other times a different way of doing something is just what the doctor ordered.

The second is treating people like numbers. This is a totalitarian perspective that runs rampant in corporate America, China, and North Korea. Businesses want measurable results and bodies in places to fill quotas. This doesn’t always lead to good results and the number one thing companies learn is how many people they’ll turn over in a year.

Micromanaging is number three. Micromanagers believe everything has to be done by the book or no one can do anything better than them. They watch their employees like hawks, creating a stifling environment. This creates stress and more mistakes on part of the employees. No one thrives with micromanagers.

The last poor management behavior is hoarding information and only sharing on a need-to-know basis if any. When this technique thrives in a work environment, employees are working in the dark with a flash light that only turns on once in a while.

Schwantes finishes with:

“Want to keep your best people? Treat them like people. Connect with them. Support them. Give them what they need to thrive. Because when you invest in your team, serve their needs, and remove obstacles from their path, they’ll run through walls for you—and that’s how you grow into the leader everyone wants to work for.”

No matter how far technology and human advances, we’ll always be dealing with these management behaviors. Dan anyone else imagine cavemen…er…people dealing with these? They probably didn’t put ump them and whacked their managers over the head with a club.

Whitney Grace, April 4, 2025

Telegram Lecture at TechnoSecurity & Digital Forensics on June 4, 2025

April 3, 2025

dino orange_thumb_thumb_thumbNo AI. Just a dinobaby sharing an observation about younger managers and their innocence.

The organizers of the June 2025 TechnoSecurity & Digital Forensics Conference posted a 60 second overview of our Telegram Overview lecture on LinkedIn. You can view the conference’s 60 second video at https://lnkd.in/eTSvpYFb. Erik and I have been doing presentations on specific Telegram subjects for law enforcement groups. Two weeks ago, we provided to the Massachusetts Association of Crime Analysts a 60-minute run down about the technical architecture of Telegram and identified three US companies providing services to Telegram. To discuss a presentation for your unit, please, message me via LinkedIn. (Plus, my son and I are working to complete our 100 page PDF notes of our examination of Telegram’s more interesting features. These range from bots which automate cross blockchain crypto movement to the automatic throttling function in the Telegram TON Virtual Machine to prevent transaction bottlenecks in complex crypto wallet obfuscations.) See you there.  — Thank you, Stephen E Arnold, April 3, 2025, 223 pm U S Eastern

Looking Busy, While Slacking

April 3, 2025

dino orange_thumb_thumb_thumb_thumbAnother dinobaby blog post. Eight decades and still thrilled when I point out foibles.

I am fascinated by people who delegate routine, courteous business functions to software and then to other people. The idea is that a busy person can accomplish much more if they are really busy but organized. I find this laughable.

In my experience, the people with full-time jobs with whom I interact are in a perpetual rush to go from one mostly pointless activity to their mobile phone and back again. Here’s an approach that has worked for some successful people. I exclude myself because I am an 80-year-old dinobaby loser.

The secret sauce consists of:

  1. Knowing what is important by day, week, month, and year. Do what’s important yourself. If you delegate, delegate with intelligence of the goal and expected outcome.
  2. Set priorities but have sufficient situational intelligence to adapt to the endlessly changing business environment. (Software just does stuff; it is not — despite the AI hype — inherently intelligent. And, no, I don’t want to discuss this perception of mine. I do not believe in made up baloney from marketing people or pressured CEOs.
  3. Recognize that how you interact with other people defines [a] your intelligence, [b] your management and social capabilities, and [c] your professional persona.

I had an email exchange a couple of days ago from a person who told me an individual would contact me. The statement was made two weeks ago. The message was, “Oh, just use our online appointment system and set up an available time.” No kidding. I am now supposed to move from “we will contact you” to navigate to our system and pick an available time. Sorry. That will not happen.

A day ago, a person who said 11 months ago, “I will call you early next week.” I received an email as it was indeed a week later, not 11 months later. Amazing. Both are considerably younger than I am, but neither person is aware of their behavior. This weird approach to business is the norm.

I read “Slack: The Art of Being Busy Without Getting Anything Done” resonated with me. I have an idea: Send the link to the article to these two people who say, “Let’s have lunch” and never call. (That’s a Manhattan trope, by the way. It means, “Hey, you, I will never call you for lunch.” Business life has become a “let’s have lunch” world. Saying something is tantamount to actually doing something.

The write up puts this in terms of a weird information sharing service which is a closed group social media thing. The write up says:

Slack brought channels and channels bought a level of almost voyeurism into what other teams were doing. I knew exactly what everyone was doing all the time, down to I knew where the marketing team liked to go for lunch. Responsiveness became the new corporate religion and I was a true believer.

To me, I think the organizations that function so that a tool like Slack is necessary have some management issues. But that’s the bias of a person who worked at a blue-chip consulting firm for longer than I thought humanly possible.

Here’s a passage I found interesting for a person paid to deliver outputs and meet objectives:

My days had become a never-ending performance of “work”. I was constantly talking about the work, planning the work, discussing the requirements of the work, and then in a truly Sisyphean twist, linking new people to old conversations where we had already discussed the work to get them up to speed on our conversation. All the while diligently monitoring my channels, a digital sentry ensuring no question went unanswered, no emoji not +1’d. That was it, that was the entire job.

What are the markers for this process of doing something that yields no deliverable that matches a job description or a task assigned by a manager to a worker?

Let me highlight a few I have noticed:

  1. Talking about doing replaces doing itself
  2. Meetings and follow ups are the work. It goes without saying that delivering an output that generates revenue is not part of the actual activity of the meeting and its follow up
  3. The mental effort required to do essentially meaningless tasks instead of satisfying deliverables of high quality burns a person out. “There is no there there.” I am not talking about Oakland, California. I am talking about the actual value to the person of doing meaningful work and getting money and mental rewards.
  4. The organization delivers increasingly degrading outputs. One wordsmith invoked feces to describe how the entity deconstructs. Microsoft shipped an update that killed its AI wunderkind Copilot. More information about new malware hit my in box today. The Epic data form for a routine visit lost the inputs I provided six months ago. My local bank charged my home checking account for over $600,000 and was unable to stop the automated fraud for two weeks.

Net net: Manage effectively and do actual work to deliver the outputs for which you are paid. Understand that both are hard. That’s why people pay you to do work. The craziness of pretending to work will make the worker crazy. If that type of person interacts with me, I just forget it. Dinobabies can do that.

Stephen E Arnold, April 3, 2025

Read AI Implements AI Into Enterprise Search

April 3, 2025

Enterprise search is an essential tool for an organization to function. There have been major shifts in enterprise search, including more accurate search results, and ability to search across multiple platforms. The Seattle startup Read AI wants to be the next enterprise search revolution says Geek Wire: “Seattle Startup Takes A Swing At Enterprise Search With New AI-Fueled Discovery Tool.”

There have been major upgrades in enterprise search before, including AI, but Read AI is doing it at level not before seen. The startup from Seattle began in 2021 with sentiment and engagement products for meetings. Later Read AI developed software tools that analyzed information across various communication channels. Now their latest product upgrade is for enterprise search that enables users to quickly locate and access information across terabytes of data and applications.

Read AI’s enterprise search tool upgrade is for Microsoft Copilot:

“Users can choose what data is discoverable. Search Copilot gets “smarter” as more workers add more data, and allows for collaboration. The new tool can also nudge users to take action based on past interactions with colleagues or customers.

The product is free to use with certain limits. Read AI CEO David Shim said reducing barriers is a key differentiator.

“People haven’t been able to understand the value of enterprise search because they haven’t had access to it,” he said.

Shim said Read AI’s ability to determine what’s important for an individual user also gives Search Copilot an edge over other AI search tools.”

Read AI wants to differentiate itself from its rivals, built-in-platform tools from Zoom, Google, and Microsoft. Read AI does have an advantage over out-of-the-box solutions, because experience tells us those tools stink. Proprietary developed tools are usually better because they’re specifically designed to enhance a specific feature, while out-of-the-box solutions are designed to with the “it’s okay to get by” mentality.

While Read AI made a smart move to upgrade Copilot with the latest AI technology, someone else would have done it eventually.

Whitney Grace, April 3, 2025

YouTube: The Future Has Blown Its Horn

April 3, 2025

YouTube has come a long way in the last two decades. Google wants us to appreciate just how far, apparently. Digiday celebrates the occasion with the piece, "As YouTube Turns 20, Here Are the Numbers You Need to Know." Writer Krystal Scanlon shares several noteworthy statistics. For example, the company states, an average of 500 hours of video is uploaded to the platform every minute. On the other end of the equation, about 30,000 viewers visit the site each day, Neilsen reported in February 2024.

In fact, we learn, YouTube considers itself the "new television." That same Neilsen report shows the platform outperforming other major streaming services. (Though it only beat Netflix by a nose, at 9.2% of total TV usage to its 8.2%.) The platform happily monetizes those eyeballs the old-fashioned way, with ads. However, it has also enticed over 125 million users to at least try its subscription plans. As for content creators, about 500,000 of them have been at it for over 10 years. They must be getting enough out of it to stick around. But for how many that is a viable career and how many it is just a hobby the write-up does not say. Whatever the case, it is clear YouTube has creators to thank for its significant ad revenue. Scanlon writes:

"With Europe’s creator economy projected to reach $41.17 billion by 2030 and U.S. influencer marketing expected to grow 14.2% to $9.29 billion this year, according to eMarketer, it’s clear why YouTube is focused on staying ahead in the creator space. Part of that plan involves doubling down on YouTube’s ability to thrust creators into the cultural mainstream. Which is why the platform has said that creators are becoming the startups of Hollywood. Whether it’s scripting, editing behind the scenes or creators are hiring various staff as part of their new business — YouTube wants to be the launchpad."

Yes, that would be quite the get. It will have to fight certain other streaming services for that honor, we think. The post continues:

"YouTube contributed more than $45 billion to the U.S. GDP in 2023, according to Oxford Economics, and it created more than 430,000 full-time jobs. Similarly, in Europe, YouTube contributed €6.4 billion ($6.94 billion) to the EU’s GDP in 2023, according to the platform’s latest Impact report, not including the jobs created by the 100,000 creators in the region that have built their own businesses including hiring staff."

Here’s the kicker. As changes roil the Google search advertising approach, YouTube might be the Little Engine That Could for Googzilla. Next up? Google Advertising, Broadcast, and Cable. ABC?

Cynthia Murrell, April 3 2025

Original Research: Not-So-Original Assertions about Content Appropriation

April 2, 2025

dino orange_thumbNo AI. Just a dinobaby sharing an observation about younger managers and their innocence.

The Social Science Research Council published the 30-plus page report “Beyond Public Access in LLM Pre-Training Data.” The subtitle reveals the principal finding: “Non-Public Non-Public Book Content in OpenAI’s Models.”

The write up states:

Using a legally obtained dataset of 34 copyrighted O’Reilly Media books, we apply the DE-COP membership inference attack method to investigate whether OpenAI’s large language models were trained on copyrighted content without consent. Our AUROC scores show that GPT-4o, OpenAI’s more recent and capable model, demonstrates strong recognition of paywalled O’Reilly book content (AUROC = 82%), compared to OpenAI’s earlier model GPT-3.5 Turbo. In contrast, GPT-3.5 Turbo shows greater relative recognition of publicly accessible O’Reilly book samples. GPT-4o Mini, as a much smaller model, shows no knowledge of public or non-public O’Reilly Media content when tested (AUROC ? 50%). Testing multiple models, with the same cutoff date, helps us account for potential language shifts over time that might bias our findings. These results highlight the urgent need for increased corporate transparency regarding pre-training data sources as a means to develop formal licensing frameworks for AI content training.

I want to mention that the DE-COP method provide one way to determine whether a specific data record was part of the training dataset for a machine learning model. The result of the SSRC’s research suggests that as OpenAI enhanced its models, the OpenAI organization appears to have used non-public content. Smaller (less capable OpenAI) models seemed to know less about the test set; the bigger models knew more.

The write up concludes that developers of smart software are sucking down data without obtaining permission or entering into licensing deals with publishers, copyright holders, or entities working on content not yet available as public content like this dinobaby blog post.

To sum up: A not-to-original assertion that certain organizations just take content and ignore rights is apparently accurate in this set of research. OpenAI accelerates. Speed is a positive. Breaking speed limits is believed to be no big deal.

A bit of a flashback: Over 40 years ago, I participated in a series of meetings to license the right to create abstracts of original articles from more than 1000 different publications. Our efforts were successful. Most publishers of academic business and management publications knew zero about online information services. The idea of summarizing an eight page article about executive compensation was unfamiliar to most of the people with whom our organization interacted over a period of years. Nevertheless, the owner of the database unit of the Courier Journal & Louisville Times Company insisted that we get signed agreements. The process was time consuming and expensive. Some outfits had representatives. So a negotiation went through layers of bureaucracy. Other publishers wanted a sit down meeting and a dog-and-pony show to explain why we wanted to boil down a 600 word article to about 150 words. There were worries about the impact of online access. There were worries about control of editorial content. There were worries about signing away rights. We couldn’t assuage worries, but we chugged along. The reputations of the Courier Journal and Barry Bingham Jr. helped but our small database operation had to follow the rules.

If the data in the cited document are accurate, the attitude of Barry Bingham Jr. would not have gained much traction at OpenAI. Instead of my getting a pat on the head for setting up the licensing deals with the able help of other Courier Journal professionals, I would have gotten kicked on the shin or worse.

Several observations:

  1. Observing content ownership remains important despite the acceleration toward illegal behaviors in the online sector
  2. Using show-and-tells and explaining the benefits of participating in digital innovations remains important. Knowledge minimizes anger and other negative reactions to pre-emptive behavior
  3. Managing the process is as important as demonstrating an unwillingness to conform to industry, social, and legal norms.

Let’s assume that the findings in the SSRC are off by 50 percent. Does that change what appears to be a strong possibility of some fancy dancing at the OpenAI disco?

My conclusion is that taking short cuts is part of the Silicon Valley approach of doing whatever and just saying, “Wow, I am sorry.”

I asked ChatGPT 4 via You.com this question:

Should I create an online product and just use content from the Harvard Business Review without permission?

Here’s the response:

No, you should not create an online product using content from the Harvard Business Review (HBR) without permission. Doing so would likely violate copyright laws and HBR’s usage policies. Here’s why:

HBR’s Copyright and Usage Policies
  1. HBR Material is Copyrighted: Harvard Business Review explicitly states that its content is protected by copyright, and no part of its publications may be reproduced without permission.
  2. Usage Restrictions: HBR has strict policies regarding how its content can be used. For example, even in academic settings, HBR articles cannot be freely distributed unless done through specific channels like creating course packs via "Harvard Business Publishing for Educators"

    Permission is Required: If you want to use HBR content, you must obtain explicit permission. This includes reproducing, distributing, or using their material in any form.

Legal and Ethical Implications
  • Copyright Infringement: Using HBR content without permission would likely constitute copyright infringement, which can result in legal action, fines, or other penalties.
  • Ethical Concerns: Beyond legal issues, using someone else’s intellectual property without authorization is unethical and could harm your reputation or business.
What You Should Do Instead
  1. Seek Permission: Contact Harvard Business Review to request permission to use their content. They may have licensing options available for your intended use.
  2. Create Original Content: Instead of using HBR’s material, consider creating your own original content inspired by their ideas (but without copying their text or proprietary materials).
  3. Use Public Domain or Licensed Content: Look for content that is in the public domain or available under a license (e.g., Creative Commons) that allows commercial use.
Conclusion

Using HBR content without permission is not only illegal but also against their explicit policies. To avoid legal and ethical issues, either seek proper licensing or create your own original material.

In this case, the OpenAI’s AI may be more intelligent that some people engaged in content acquisition at the firm. But I am a dinobaby. I try to follow the social, legal, and ethical rules as my parents who had only high school educations taught me. I am a dinobaby. A sad dinobaby.

Stephen E Arnold, April 2, 2025

A Reminder That Academic Papers Can Be Science Fiction

April 2, 2025

dino orange_thumbDinobaby says, “No smart software involved. That’s for “real” journalists and pundits.

For many years, I have been skeptical about peer reviewed papers. There are two reasons.

First, I did some consulting work for an outfit in the UK. It was a crown operation. That means its outputs carried considerable influence. One of the projects in which I was engaged involved peer review of research under the purview of the “institute.” Wow, did I learn why there was a publishing backlog. Wow, did I learn about the cronyism in reviewing “real” research. Wow, did I learn about the wheeling and dealing of peer reviewers among their research assistants. Wowzah. That was an education.

Second, for a short time I was a peer reviewer for a British journal. Let me tell you that my first hand exposure to the mechanics and politics of peer reviewing did not prepare me for the reviewing task. A typical submission contained text edited by several hands. None of these was doing fine needlework. A stab here and a stab these summed up the submitted documents. The data and the charts? I had a couple of my team help me figure out if the chart was semi accurate. Working through a five or six page article sent to me for review took me and two people a week to process. In most cases, we gave the paper a D and sent it back to the editor in chief who had to tell the author and his legion of busy bees that the paper sucked. I bailed after six months. Too much work to fix up stuff that was truly terrible.

Today I read “Sometimes Papers Contain Obvious Lies.” That’s a good title, but my thought would be to include the phrase “and Really Crappy.” But I am a dinobaby, and I live in rural Kentucky. The author Cremieux Recueil is much classier than I.

I noted this passage:

The authors of scientific papers often say one thing and find another; they concoct a story around a set of findings that they might not have even made, or which they might have actually even contradicted. This happens surprisingly often, and it’s a very serious issue…

No kidding. The president of Stanford University resigned due to some allegations of fancy dancing. The — note the the — Harvard University experienced a bit of excitement in its ethics department. Is that an oxymoron? An ethics professors violated “ethics” in some research cartwheels.

I liked this sentence because it is closer to my method of communicating concern:

Lying in scientific papers happens all the time.

Hey, not just in scientific papers. I encounter lying 24×7. If someone is not articulating a fabrication, the person may be living a lie. I hear the roar of a 20 somethings hyper car at the gym. Do you?

The paper focuses on a paper with some razzle dazzle related to crime data. The author’s analysis is accurate. However, the focus on an example does not put the scale of the “crime data” problem in perspective.

Let me give you an example and you can test this for validity yourself. Go to your bank. Ask the “personal banker” to tell you about the bank’s experience with cyber crime. Then ask, “How many fraudulent transactions occur at this bank location each year?” Listen to the answer.

Crime data, like health care data, are slippery fish. Numbers do not correlate to reality when scrutinized. Verifiable, statistically valid data is expensive to generate. We live in a “good enough” world and trust whatever black box (human or mechanical) spits out data.

I do disagree with this statement in the essay:

scientists often lie with far more brazenness.

No. Fabrication is now the business of information and the information of business.

Stephen E Arnold, April 2, 2025

The AI Market: The Less-Educated

April 2, 2025

Writing is an essential function of education and communication. Writing is an innate skill as well as one that can be curated through dedicated practice. Digital writing tools such as spelling and grammar checkers and now AI like Grammarly and ChatGPT have influenced writing. Stanford University studied how AI writing tools have impacted writing in professional industries. The discovered that less-educated parts of the US heavily rely on AI. Ars Technica reviews the study in: “Researchers Surprised To Find Less-Educated Areas Adopting AI Writing Tools Faster.”

Stanford’s AI study tracked LLM adoption from January 2022 to September 2024 with a dataset that included US Consumer Financial Protection Bureau consumer complaints, corporate press releases, job postings, and UN press releases. The researchers used a statistical detection system that tracked word usage patterns. The system found that 14-24% of these communications showed AI assistance. The study also found an interesting pattern:

“The study also found that while urban areas showed higher adoption overall (18.2 percent versus 10.9 percent in rural areas), regions with lower educational attainment used AI writing tools more frequently (19.9 percent compared to 17.4 percent in higher-education areas). The researchers note that this contradicts typical technology adoption patterns where more educated populations adopt new tools fastest.”

The researchers theorize that AI-writing tools serve as equalizing measures for less-educated individuals. They also noted that AI-writing tools are being adopted because the market is saturated or the LLMs are becoming more advanced. IT will be difficult to distinguish between human and machine written text. They predict negative outcomes from this:

“ ‘The growing reliance on AI-generated content may introduce challenges in communication,’ the researchers write. ‘In sensitive categories, over-reliance on AI could result in messages that fail to address concerns or overall release less credible information externally. Over-reliance on AI could also introduce public mistrust in the authenticity of messages sent by firms.’”

It’s not good to blindly trust AI, especially with the current state of datasets. Can you imagine the critical thinking skills these future leaders and entrepreneurs will develop? On that thought, what will happen to imagination?

Whitney Grace, April 2, 2025

The First AI-Written Paper To Pass Peer Review

April 2, 2025

Cheating. I am not going to bring this topic up.

Humans have taken one small stop towards obsolesce when it comes to writing papers. Sakana AI reports that "The AI Scientist Generates Its First Peer-Reviewed Scientific Publication." This is the first known fully AI-generated paper that passed the same review process that human scientists submit their papers too. Here’s how the paper was written:

"The paper was generated by an improved version of the original AI Scientist, called The AI Scientist-v2. We will be sharing the full details of The AI Scientist-v2 in an upcoming release. This paper was submitted to an ICLR 2025 workshop that agreed to work with our team to conduct an experiment to double-blind review AI-generated manuscripts. We selected this workshop because of its broader scope, challenging researchers (and our AI Scientist) to tackle diverse research topics that address practical limitations of deep learning. The workshop is hosted at ICLR, one of three premier conferences in machine learning and artificial intelligence research, along with NeurIPS and ICML.3

The ICLR leadership and organizers were involved with the project. The paper was blindly submitted to the ICLR review team, although they were told that they might be reviewing AI generated papers.

The AI algorithm was told to research and write about a broad topic. When the process was done, three papers were selected for submission so the review board wouldn’t be overburdened. Here are the results:

“We looked at the generated papers and submitted those we thought were the top 3 (factoring in diversity and quality—We conducted our own detailed analysis of the 3 papers, please read on in our analysis section). Of the 3 papers submitted, two papers did not meet the bar for acceptance. One paper received an average score of 6.33, ranking approximately 45% of all submissions. These scores are higher than many other accepted human-written papers at the workshop, placing the paper above the average acceptance threshold. Specifically, the scores were:

• Rating: 6: Marginally above acceptance threshold

• Rating: 7: Good paper, accept

• Rating: 6: Marginally above acceptance threshold”

The AI Scientist conducted the experiment out of pure scientific curiosity to measure how current AI algorithms compare to human intellect. No problem.

Whitney Grace, April 2, 2025

No Joke: Real Secrecy and Paranoia Are Needed Again

April 1, 2025

dino orangeNo AI. Just a dinobaby sharing an observation about younger managers and their innocence.

In the US and the UK, secrecy and paranoia are chic again. The BBC reported “GCHQ Worker Admits Taking top Secret Data Home.” Ah, a Booz Allen / Snowden type story? The BBC reports:

The court heard that Arshad took his work mobile into a top secret GCHQ area and connected it to work station. He then transferred sensitive data from a secure, top secret computer to the phone before taking it home, it was claimed. Arshad then transferred the data from the phone to a hard drive connected to his personal home computer.

Mr. Snowden used a USB drive. The question is, “What are the bosses doing? Who is watching the logs? Who is  checking the video feeds? Who is hiring individuals with some inner need to steal classified information?

But outside phones in a top secret meeting? That sounds like a great idea. I attended a meeting held by a local government agency, and phones and weapons were put in little steel boxes. This outfit was no GHCQ, but the security fellow (a former Marine) knew what he was doing for that local government agency.

A related story addresses paranoia, a mental characteristic which is getting more and more popular among some big dogs.

CNBC reported an interesting approach to staff trust. “Anthropic Announces Updates on Security Safeguards for Its AI Models” reports:

In an earlier version of its responsible scaling policy, Anthropic said it would begin sweeping physical offices for hidden devices as part of a ramped-up security effort.

The most recent update to the firm’s security safeguards adds:

updates to the “responsible scaling” policy for its AI, including defining which of its model safety levels are powerful enough to need additional security safeguards.

The actual explanation is a master piece of clarity. Here’s snippet of what Anthropic actually said in its “Anthropic’s Responsible Scaling Policy” announcement:

The current iteration of our RSP (version 2.1) reflects minor updates clarifying which Capability Thresholds would require enhanced safeguards beyond our current ASL-3 standards.

The Anthropic methods, it seems to me, to include “sweeps” and “compartmentalization.”

Thus, we have two examples of outstanding management:

First, the BBC report implies that personal computing devices can plug in and receive classified information.

And:

Second, CNBC explains that sweeps are not enough. Compartmentalization of systems and methods puts in “cells” who can do what and how.

Andy Grove’s observation popped into my mind. He allegedly rattled off this statement:

Success breeds complacency. Complacency breeds failure. Only the paranoid survive.

Net net: Cyber security is easier to “trust” and “assume”. Real fixes edge into fear and paranoia.

Stephen E Arnold, April 9, 2025

« Previous PageNext Page »

  • Archives

  • Recent Posts

  • Meta