5 Filters

Minibrained's latest cabon sequestration lunacy

Hi folks I noticed this on WUWT website:

“In areas with any human or animal habitation, large sudden releases are potentially fatal since heavier-than-air CO2 can drop like a stone, drive out oxygen and asphyxiate anyone beneath. One line of enquiry, common when such incidents occur, was that the rupture was caused by the formation of carbonic acid. This is a major problem in CO2 pipelines since the acid can form with trace amounts of water – as little as 100 parts per million. Just one further hazard to consider as the hard-Left Miliband lunatics press ahead with billion-pound plans to run hundreds of miles of near-surface, large-diameter pipes around the north of England to bury CO2 in quantities that will have no measurable effect on any change in the climate.”

The whole CO2 hoax has been captured by the AI hoax - an amusing article here:
" Peter Girnus :eagle:@gothburz

15h

Last year the board created a new position. Chief AI Officer. They gave it to me. I had no qualifications. I had a LinkedIn post about prompt engineering that got 4,000 likes. Someone on the nominating committee saw it. They said I had “a demonstrated fluency in the space.” The space. Nobody defines the space. The space is whatever the quarterly earnings call needs it to be. My first act was to commission a roadmap. McKinsey built it. $4.2 million. Fourteen weeks. The roadmap recommended building a Center of Excellence. The Center of Excellence’s first deliverable was a second roadmap. I approved both roadmaps. The second roadmap recommended retaining Accenture. Accenture recommended retaining BCG. BCG recommended a “maturity assessment.” The maturity assessment costs $1.8 million and measures how mature your AI program is on a scale from 1 to 5. We scored a 1.4. BCG said this was “expected for organizations at the early stages of their transformation journey.” The transformation journey. The journey costs $480 an hour. The tools cost $30 a month. I pay sixteen times more for the people explaining the tools than I pay for the tools. This is called “strategic implementation.” I built the Center of Excellence. Eleven people. A director, two program managers, three analysts, a communications specialist, a change management lead, two vendor liaisons, and an AI Ethics Advisor. The AI Ethics Advisor’s job is to attend meetings where no one asks ethical questions. She has attended all of them. The Center of Excellence’s primary output is a biweekly newsletter called “AI Horizons.” It goes to all 14,000 employees. The open rate is 6%. I know this because the analytics dashboard for the newsletter is the most functional AI tool we’ve deployed. I reported to the board that we had “reached 14,000 stakeholders with our AI literacy initiative.” That is technically true. Reaching is not the same as reading. But reaching is a metric. Reading is not. In September I rolled out an AI-powered meeting summarization tool. It replaced Janet. Janet took notes by hand. Janet was accurate. Janet cost $58,000 a year. The AI tool costs $340,000 a year in licensing, integration, and cloud compute. It summarized a board meeting by attributing the CFO’s layoff announcement to the head of diversity. It summarized a product review by recommending we “sunset the customer.” We have not sunsetted any customers. We sunsetted Janet. Janet was a contractor. Contractors don’t count in the headcount reduction. She counts in the AI displacement metric. I created that metric. It goes up and to the right. In October I launched AskHR, an AI chatbot for employee questions. It was supposed to reduce HR ticket volume by 40%. In the first week, it told an employee his paternity leave was fourteen months. It was fourteen days. The employee told his wife. His wife told her employer. HR told me. I told the board the chatbot had “achieved a 40% deflection rate.” Deflection means the employee stopped asking. It does not mean the employee got an answer. Forty percent of people who interact with AskHR give up. I call this deflection. Customer Success calls this resolution. The dashboard calls this a closed ticket. The employee calls it nothing, because nobody asks the employee. In November, Goldman Sachs published a report. It said AI had contributed “basically zero” to GDP despite $650 billion in infrastructure spending. Basically zero. I forwarded the report to the CEO with the subject line: “This is why we need to ACCELERATE.” He agreed. We accelerated. We are now spending $28 million annually on AI initiatives. The initiatives are: one newsletter with a 6% open rate, one chatbot that hallucinates paternity leave, one meeting summarizer that recommends sunsetting customers, one AI Ethics Advisor who has never been asked an ethical question, and one maturity assessment that says we’re a 1.4. Only 25% of enterprises report that generative AI has transformed their business. I know this because Deloitte surveyed 3,235 leaders and put it in a PDF. I put the PDF in the board deck. I highlighted the 25% and wrote “massive upside remaining.” The board loved that. Massive upside remaining means we haven’t seen results yet and we’re choosing to frame that as an opportunity rather than an indictment. Only 25% believe their IT infrastructure supports enterprise-wide AI. I know this because IBM surveyed 400 leaders and put it in a different PDF. I put both PDFs in the same slide. I titled the slide “The Readiness Gap.” I did not title it “We Are Not Ready.” Forty-four percent of organizations have no measurement framework for generative AI. I am in the 44%. I have a dashboard. A dashboard is not a measurement framework. A dashboard is a picture of numbers. The numbers go up because I designed them to go up. Adoption is measured by logins. A login is counted when the application opens. It opens automatically when you turn on your laptop. Every employee in the company logs in to the AI platform every morning. None of them meant to. I reported this as “98% daily active engagement.” The board gave me a standing ovation. Standing ovations are not a measurement framework either but they are the only metric that determines my bonus. In February, OpenAI’s COO said — and I need you to hear this — “we have not yet really seen AI penetrate enterprise business processes.” He said this out loud. At a conference. He runs OpenAI’s enterprise division. He said the product has not penetrated the market he is selling it to. The next day, OpenAI announced “Frontier Alliances” — multi-year partnerships with McKinsey, BCG, Accenture, and Capgemini. The AI company hired the consultants. The consultants had already been hired by us to implement the AI company’s product. The AI company is now paying the consultants to help the consultants help us use the AI that the AI company built. This is not a loop. This is a business model. The business model has a name. The name is “ecosystem.” BCG’s CEO said, and this is a direct quote: “AI alone does not drive transformation. It must be linked to strategy.” Strategy is the thing BCG sells. The same week, Anthropic’s head of product said: “2025 was meant to be the year agents transformed the enterprise, but the hype turned out to be mostly premature.” Premature. That is the word you use when you want to say “we were wrong” without accepting that anyone was harmed. 108,435 people lost their jobs in January. That’s the highest monthly total since 2009. Some of those people lost their jobs because companies like mine said AI would replace them. The AI has not replaced them. The AI has replaced their budget line. Their work is now done by the remaining employees, who are doing two jobs for one salary, which we call “efficiency gains.” Efficiency gains go in the board deck. The remaining employees got raises this year. Flat raises. Same percentage for everyone. The industry calls this “peanut butter.” You spread it thin and even so no one complains loudly enough to matter. Forty-four percent of companies are doing peanut butter raises this year. It is the same tactic we used after the 2008 recession. We are running the 2008 playbook while claiming to be in the future. I attended the World Economic Forum in January. I was on a panel called “Responsible AI at Scale.” There were four of us on the panel. None of us could define what our AI platforms measurably do. All of us used the word “transformative.” One of us used it four times in a single answer. That one was me. I have created an AI Readiness Score. The score is calculated by multiplying the number of AI tools licensed by the number of departments that have opened them divided by a normalization factor I derived from a Harvard Business Review article I skimmed. Our readiness score is 74 out of 100. Seventy-four means we are “Advanced — Approaching Optimization.” Approaching means we haven’t arrived. Optimization means nothing. The score is on a poster in the lobby of Building 3. It is printed in Helvetica Neue 120pt. Beneath the score it says “Powered by People, Accelerated by AI.” I wrote that tagline. It won an internal communications award. The award is called the Catalyst Prize. I also created the Catalyst Prize. The maturity model has five levels. Level 5 is “AI-Native Organization.” Nobody has reached Level 5. The assessment firm told me this privately. They also told me the model was designed so that reaching Level 5 requires re-engaging the assessment firm for “advanced optimization services.” The model is a subscription. The destination is unreachable. I presented the annual AI strategy to the board last week. Total spend: $28 million. Measurable productivity gain: pending. Pending means we haven’t measured it and we’ve decided that not measuring it is a strategy rather than an omission. The board approved a $42 million budget for next year. They approved it in nine minutes. Nobody asked what the $28 million accomplished because I showed them the readiness score and the readiness score went from 58 to 74 and 74 is bigger than 58 and bigger is better and better means we’re winning. I will be promoted to EVP by Q3. My department has produced one newsletter, one broken chatbot, one meeting tool that recommends sunsetting customers, one ethics advisor who advises no one, two roadmaps, three consulting engagements, a maturity score of 1.4, a readiness score of 74, and a poster in Helvetica Neue. Goldman Sachs says the impact is basically zero. My compensation says otherwise. The graph goes up and to the right. It measures AI investment. Investment means spending. Spending means commitment. Commitment means we’re serious about the future. The future is whatever I say it is. As long as the graph goes up and to the right."

cheers

1 Like