All Activity
- Past hour
-
Google adds channel performance timeline view to PMax campaigns
Google launched a new channel performance timeline view inside Performance Max, giving advertisers a clearer breakdown of how individual channels — Search, YouTube, Display, and others — are contributing to campaign results over time. What’s new. A timeline graph now shows channel-level contributions over a selected period, paired with investment and performance filters. Advertisers can see at a glance which channels are pulling their weight — and which aren’t. Yellow box – Channel Performance Evolution Over Time Pink box (right) – All Ads, Ads Using Product Lists, Ads Using Video Why we care. Performance Max campaigns span multiple channels simultaneously, which makes it genuinely difficult to know where budget is being spent most effectively. This update gives advertisers a timeline view of channel-level contributions — so if YouTube is underperforming while Search is driving most conversions, that’s now visible without having to dig through export data or rely on guesswork. This allows you to spot channel-level trends earlier and adjust asset strategy or budget accordingly. The big picture. The view gives advertisers a more actionable way to evaluate PMAX performance without having to rely solely on Google’s automated decisions. If YouTube is consistently underdelivering while Search drives the bulk of conversions, that’s now visible — and can inform budget and asset strategy going forward. The bottom line. It’s not full transparency, but it’s a meaningful step in the right direction. Advertisers have a cleaner way to spot PMax trend anomalies early and adjust accordingly. First spotted. This update was first spotted by Head of Search at Le Mage du SEA, Axel Falck, who shared spotting this update on LinkedIn. View the full article
-
This High-Capacity Portable Power Station Is $500 Off Right Now
We may earn a commission from links on this page. Deal pricing and availability subject to change after time of publication. DJI built its reputation on drones and creator gear, but the brand has been making serious moves in the portable power space. The DJI Power 2000 is the company's most capable standalone power station yet. Think of it as a large rechargeable battery you can take anywhere—camping, a job site, an outdoor movie night, or just kept in the garage for when the power goes out. At $799, it's currently down $500 from its original $1,299 price tag. That's a solid discount for something that sits in the higher end of this category. DJI Power 2000 Portable Power Station Electric generator for home, camping & RVs $799.00 at Amazon $1,299.00 Save $500.00 Get Deal Get Deal $799.00 at Amazon $1,299.00 Save $500.00 It's about the size of a carry-on suitcase and weighs close to 50 pounds, so it's not something you'd strap to your back, but it fits easily in a car trunk. The battery itself is built to last, rated for 4,000 charges before it starts losing any noticeable capacity. Performance-wise, DJI says the 2,048Wh capacity can keep a refrigerator running for up to 40 hours during a home power outage, charge a laptop 18 times, or power a projector for 18 hours. Those numbers line up with what you’d expect from a unit this size, and they make it useful for both emergencies and planned outdoor use. You also get a solid mix of ports—there are three AC outlets, a 30-amp port for heavier appliances or RV setups, multiple USB-C ports (including a 140W option for laptops), and DJI’s own SDC ports for direct drone charging. If you use DJI gear, that alone makes life easier in the field, notes this Mashable review. Recharging is also fast for a battery this size. Plugged into a wall outlet, it can go from empty to full in about 90 minutes. You can also top it up with solar panels or from your car, which makes it workable for longer trips. Rounding out its features is its companion app, which lets you check its battery level from your phone without having to walk over to it, a small thing that turns out to be pretty convenient. Note that this unit has no built-in light, which is something many competing power stations include. Our Best Editor-Vetted Tech Deals Right Now Apple AirPods Pro 3 Noise Cancelling Heart Rate Wireless Earbuds — $199.00 (List Price $249.00) Apple iPad 11" 128GB A16 WiFi Tablet (Blue, 2025) — $299.00 (List Price $349.00) Samsung Galaxy Tab A11+ 128GB Wi-Fi 11" Tablet (Gray) — $202.00 (List Price $249.99) Apple Watch Series 11 (GPS, 42mm, S/M Black Sport Band) — $329.00 (List Price $399.00) Sony WH-1000XM5 — $298.00 (List Price $399.99) Deals are selected by our commerce team View the full article
-
Globalstar stock soars on Amazon rumors. Why the Starlink rival’s shares are blasting into space today
Most of the markets are down today after President The President’s address to the nation last night failed to alleviate fears about America’s war with Iran dragging on. But one relatively small tech company is bucking the downward trend in premarket trading this morning: Globalstar. The satellite communications company is reportedly an acquisition target for Amazon, yet its relationship with Apple could complicate any potential deal. Here’s what you need to know. What’s happened? Shares in the relatively small satellite communications company Globalstar, Inc. (Nasdaq: GSAT) are rising today after a Financial Times report yesterday said the ecommerce giant Amazon was in talks to acquire the company. As of the time of this writing, GSAT shares are up nearly 14% to around $78 apiece—making it one of the only tech companies to see significant gains this morning in premarket trading. The company’s share price closed yesterday at $68.53, putting its valuation at around $9 billion. While Globalstar is not a household name, the company has been in operation for over 35 years. It is the operator of a network of low Earth orbit satellites that provide communication services to companies across industries, from defense to Big Tech. Why does Amazon want to buy Globalstar? It’s important to note that while Amazon is rumored to be in negotiations with Globalstar, a deal could still fail to materialize, according to the FT. That being said, the obvious main driver behind Amazon’s interest in Globalstar would be to gain access to its network of satellites so Amazon can better compete with Elon Musk’s Starlink satellite internet service, which is operated by Musk’s company, SpaceX. Amazon is currently building out its competing high-speed satellite internet service, Leo, which aims to challenge Starlink’s dominance in the space. However, Leo currently lags significantly behind Starlink in the number of satellites each company operates. While SpaceX has over 10,000 satellites in its Starlink network, Amazon’s Leo has a constellation of less than 200. By acquiring Globalstar, Amazon could play catchup not just by adding the company’s satellites to its existing fleet, but by acquiring the technology, talent, and expertise that the 35-year-old company possesses. That’s something Amazon needs as it expands its Leo service. Yesterday, the company announced that Leo would power in-flight Wi-Fi for Delta Air Lines in 2028. Apple’s relationship with Globalstar could complicate things Unfortunately for Amazon, if it wants to acquire Globalstar, it doesn’t only need to negotiate with the satellite provider itself. Amazon also needs to deal with Apple. That’s because Apple actually owns 20% of Globalstar. The iPhone giant made a $1.1 billion investment in the satellite firm in 2024, giving it partial ownership of the company as part of the deal. As 9to5Mac reported at the time, Apple made the investment to help Globalstar expand its infrastructure, which Apple uses to power its “Emergency SOS” features on the iPhone and Apple Watch. The feature allows iPhone and Apple Watch users to make emergency phone calls and send emergency texts from their devices, even when they have no standard cellular signal. This is possible due to the satellite connectivity Globalstar provides. The Financial Times reported that Apple’s 20% ownership of Globalstar was a “complicating factor” in the current negotiations. Fast Company has reached out to Amazon, Apple, and Globalstar for comment. GSAT stock jumps while AAPL and AMZN fall It’s little surprise that GSAT is surging on rumors that the deep-pocketed Amazon was in talks to acquire Globalstar. As of this writing, it is up 14% in premarket trading. That’s on top of yesterday’s 3% gain for GSAT. And even before today’s premarket price jump of nearly 14%, GSAT shares have been on a roll. As of yesterday’s close, GSAT shares were up more than 12% for the year, and over the past 12 months, they have risen an astounding 230%. Globalstar’s stock price gain contrasts starkly with the fall in Apple Inc. (AAPL) and Amazon.com, Inc. (AMZN) shares today. The stock prices of both tech giants are down in the low single digits as of the time of this writing. But those price declines likely are a reflection of the broader market selloff after The President’s address to the nation on Wednesday, which did not alleviate fears about the Iran war dragging on. And as for Starlink, Reuters reported yesterday that SpaceX, the Amazon Leo competitor’s parent company, has confidentially filed for its initial public offering, which is expected in June. View the full article
- Today
-
AA attracts interest from EQT as it prepares for £5bn sale
UK roadside recovery business attracts multiple suitors as owners seek exitView the full article
-
How To Identify And Solve Click Fraud In Paid Media – Ask A PPC via @sejournal, @navahf
This week's Ask A PPC dives into how to identify and solve click fraud in paid media to optimize your advertising efforts. The post How To Identify And Solve Click Fraud In Paid Media – Ask A PPC appeared first on Search Engine Journal. View the full article
-
What Are Stabilizer Muscles (and Do You Really Need to Train Them)?
You may have heard dumbbell exercises are better than barbell ones because they work more of your “stabilizers,” or that free weights are better than machines for the same reason. But what are stabilizer muscles? And do you really need specific exercises to train them? It turns out there are a lot of misconceptions around this term, so let me set things straight. What are stabilizer muscles?This is going to get fuzzy, because there isn’t really agreement on what stabilizer muscles even are. This 2014 study searched the literature for mentions of stabilizer muscles and attempted to put together a definition. Here’s what they came up with: "muscles that contribute to joint stiffness by co-contraction and show an early onset of activation in response to perturbation via either a feed-forward or a feedback control mechanism." Okay, stabilizer muscles are muscles that, well, stabilize. Which muscles are those? That’s a harder question. A muscle might stabilize a joint while doing one type of exercise or motion, but that doesn't mean it always acts as a stabilizer. Just as an actor can play a supporting role in one movie and a starring role in another, muscles don't have to be limited to just a "stabilizer" role. Stabilizers thus aren't a type of muscle, but rather a role that a muscle may or may not play in a given context. Taking this back to the scientific literature: you can find plenty of research on “lumbar [lower back] stabilizers” or “trunk [core] stabilizers” or “knee stabilizers.” But these don’t turn out to be specific muscles that only stabilize joints. This study on knee stabilizers names four muscles that are part of the quadriceps and hamstring muscle groups (the big muscle groups on the front and back of the thigh, respectively). Are those stabilizers, or are they simply muscles that move the legs? One exercise’s stabilizers may be another’s main moversThis is why I don’t worry too much about a certain exercise routine (say, one that sticks to weight machines) neglecting “stabilizing” muscles. If you do a variety of quad exercises and a variety of hamstring exercises, you’re pretty much guaranteed to hit the quad and hamstring muscles that act as knee stabilizers when you’re running and jumping. Or to use another example: Single-leg exercises like step-ups and lunges are great for working your abductors (hip muscles) and adductors (inner thigh muscles) because those muscles work to keep your leg steady as you put weight on it. But if a person never did single leg exercises, they could still hit those muscles by doing exercises that target them as main movers, like the adductor and abductor machines. Being stable is about coordination, not just strengthIf we look again at research on knee stabilizers, scientists have a theory that it’s good for injury prevention if your body uses those stabilizer muscles while running and jumping. This isn’t just about the strength of those muscles, but also your ability to activate them when they’re needed. So the way you keep your knees stable is not just by doing free weight exercises—although those are great—but also by doing running, jumping, pivoting, and cutting exercises. (Think soccer players running around cones and rope ladders.) In other words, practice is important to joint stability, not just strength. If you want to be steady and stable while performing certain motions, you’ll need to train your brain to drive those muscles at the right time and in the right order. Strength and stability are sometimes at oddsSo what should you do in the gym? You may notice that strong people usually train with a mix of exercises. They might squat and bench with a barbell, but finish off their sessions with a dumbbell bench press or the leg extension machine. There is a continuum to working out, with strength on one end and stability on the other, and each of those exercises falls at a different point on that continuum. Let’s use bench press as our example. In a barbell bench press, you need to use your legs to stabilize your torso, your torso to make a stable platform for your arms, and your arms to move the weight. Even though you’re training your pecs and triceps as the main movers, you’re getting a lot of shoulder, core, back, and leg muscles involved as stabilizers. We can involve our stabilizers more if we were to do something like a dumbbell bench press with our back on a yoga ball. We would have to work harder to keep everything steady, but as a result, we wouldn’t be able to use nearly as much weight. We would be training stabilizers more but the main movers (chest and triceps) less. The opposite end of that spectrum would be a chest press machine. There, you don’t have to do much stabilizing at all—just whatever it takes to sit in the chair without falling out. The pecs and triceps are no longer limited by what our stabilizers can handle, so we can “lift” even more weight. (That of course comes with the caveat that you can’t compare machine labels to barbell or dumbbell weights; the mechanics are different.) So do you need to “train” your stabilizers?My take is this: If you train every part of your body, no matter how you do it, you will end up training all your stabilizer muscles. Yes, even if you do an all-machine routine. The routine only has to be well-rounded. If you’ve been sticking with “functional” exercises that require a lot of stabilization, you are probably doing plenty for your stabilizers without really thinking about it. The tradeoff is that you may not be giving the main movers of each exercise as much work. You can easily get the best of both worlds by doing a variety of exercises. If you never do anything that makes you feel unstable, add some single-leg exercises, carries, or other slightly unstable work to your routine. (No need to stand on a bosu, although you can if you want, I guess.) And if you do a lot of stability work, try out some machines or barbell exercises once in a while to make sure you’re building strength too. View the full article
-
Build your own AI search visibility tracker for under $100/month
Tracking your brand’s visibility in AI-powered search is the new frontier of SEO. The tools built to do this are expensive, often starting at $300 to $500 per month and quickly rising from there. For many, that price is a nonstarter, especially when custom testing needs go beyond what off-the-shelf software can handle. I faced this exact problem. I needed a specific tool, and it didn’t exist at a price I could afford, so I decided to build it myself. I’m not a developer. I spent a weekend talking to an AI agent in plain English, and the result was a working AI search visibility tracker that does exactly what I need. Below is the guide I wish I’d had when I started: a step-by-step playbook for building your own custom tool, covering the technology, the process, what broke, and how to get it right faster. The problem: A custom tool for a complex landscape My goal was to automate an AI engine optimization (AEO) testing protocol. This wasn’t just about checking one or two models. To get a full picture of AI-driven brand visibility, I knew from the start that we had to track five distinct, critical surfaces: ChatGPT (via API): The most well-known conversational AI. Claude (via API): A major competitor with a different response style. Gemini (via API): Google’s direct, developer-facing model. Google AI Mode: Google’s AI search experience, which uses Gemini 3 for advanced reasoning and multimodal understanding. Google AI Overviews: The summary boxes that appear at the very top of the SERP for many queries, which by late 2025 were appearing in nearly 16% of all Google searches. On top of that, I needed to score the results using a custom 5-point rubric: brand name inclusion, accuracy, correctness of pricing, actionability, and quality of citations. No existing SaaS tool offered this exact combination of surfaces and custom scoring. The only path forward was to build. Here are a few screenshots of the internal tool as it stands. You can see some of my frustration in the agent chat window. The method: Using vibe coding to build the tool This project was built using vibe coding, a way of turning natural language instructions into a working application with an AI agent. You focus on the goal, the “vibe,” and the AI handles the complex code. This isn’t a fringe concept. With 84% of developers now using AI coding tools and a quarter of Y Combinator’s Winter 2025 startups being built with 95% AI-generated code, this method has become a viable way for non-developers to create powerful internal tools. Dig deeper: How vibe coding is changing search marketing workflows Your customers search everywhere. Make sure your brand shows up. The SEO toolkit you know, plus the AI visibility data you need. Start Free Trial Get started with Your tech stack: The three tools you’ll need You can replicate this entire project with just three things, keeping your monthly cost under $100. Replit Agent This is a development environment that lives entirely in your web browser. Its AI agent lets you build and deploy applications just by describing what you want. You don’t need to install anything on your computer. The plan I used costs $20/month. DataForSEO APIs This was the backbone of the project. Their APIs let you pull data from all the different AI surfaces through a single, unified system. You can get responses from models like ChatGPT and Claude, and pull the specific results from Google’s AI Mode and AI Overviews. It has pay-as-you-go pricing, so you only pay for what you use. Direct LLM APIs (optional but recommended) I also set up direct connections to the APIs for OpenAI (ChatGPT), Anthropic (Claude), and Google (Gemini). This was useful for double-checking results and debugging when something seemed off. The playbook: A step-by-step guide to building your tool Building with an AI agent is a partnership. The AI will only do what you ask, so your job is to be a clear and effective guide. Here’s a repeatable framework that will help you avoid the biggest mistakes. Step 1: Write a requirements document first Before you even open Replit, create a simple text document that outlines exactly what you need. This is your blueprint. Include: The core problem you’re solving. Every feature you want (e.g., CSV upload, custom scoring, data export). The data you’ll put in, and the reports you want out. Any APIs you know you’ll need to connect to. Start your conversation with the AI agent by uploading this document. It will serve as the foundation for the entire build. Step 2: Ask the AI, ‘What am I missing?’ This is the most important step. After you provide your requirements, the AI has context. Now, ask it to find the blind spots. Use these exact questions: “What am I not accounting for in this plan?” “What technical issues should I know about?” “How should data be stored so my results don’t disappear?” That last question is critical. I didn’t ask it, and I lost a whole batch of test results because the agent hadn’t built a database to save them. Step 3: Build one feature at a time and test it Don’t ask the AI to build everything at once. Give it one small task, like “build a screen where I can upload a CSV file of prompts.” Once the agent says it’s done, test that single feature. Does it work? Great. Now move to the next one. This incremental approach makes it much easier to find and fix problems. Dig deeper: How to vibe-code an SEO tool without losing control of your LLM Get the newsletter search marketers rely on. See terms. Step 4: Point the agent to the documentation When it’s time to connect to an API like DataForSEO, don’t assume the AI knows how it works. Find the API documentation page for what you’re trying to do, and give the URL directly to the agent. A simple instruction like, “Read the documentation at this URL to implement the authentication,” will save you hours of frustration. My first attempt at connecting failed because the agent guessed the wrong method. Step 5: Save working versions Before you ask for a major new feature, save a copy of your project. In Replit, this is called “forking.” New features can sometimes break old ones. I learned this when the agent was working on my results table, and it accidentally broke the CSV upload feature that had been working perfectly. Having a saved version makes it easy to go back and see what changed. Dig deeper: Inspiring examples of responsible and realistic vibe coding for SEO What will break: A field guide to common problems Nearly everything will break at some point. That’s part of the process. Here are the most common issues I ran into, and the lessons I learned, so you can be prepared. ProblemThe lesson and how to fix it1. API authentication failsThe agent will often try a generic method. Fix: Give the agent the exact URL to the API’s authentication documentation.2. Results disappearThe agent may not build a database by default, storing data in temporary memory instead. Fix: In your first step, ask the agent to include a database for persistent storage.3. API responses don’t show upYou might see data in your API provider’s dashboard, but it’s missing in your app. This is usually a parsing error. Fix: Copy the raw JSON response from your API provider, and paste it into the chat. Say, “The app isn’t displaying this data. Find the error in the parsing logic.”4. Model responses are cut shortAn LLM like Claude might suddenly start giving one-word answers. This often means the token limit was accidentally changed. Fix: After any update, run a quick test on all your connected AI surfaces to ensure the basic parameters haven’t changed.5. API results don’t match the public versionChatGPT’s public website provides web citations, but the API might not. Fix: Realize that APIs often have different default settings. You may need to explicitly tell the agent to enable features like web search for the API call.6. Citation URLs are unusableGemini’s API returned long, encoded redirect links instead of the final source URLs. Fix: Inspect the raw data. You may need to ask the agent to build a post-processing step, like a redirect resolver, to clean up the data.7. Your app isn’t updatedYou build a great new feature, but it doesn’t seem to be working in the live app. Fix: Understand the difference between your development environment and your production app. You need to explicitly “publish” or “deploy” your changes to make them live. The real costs: Is it worth it? Building this tool saved me a significant amount of money. Here’s a simple cost comparison against a mid-tier SaaS tool. ItemDIY tool (My project)SaaS alternativeSoftware subscription~$20/month (Replit)$500/monthAPI usage~$60/month (variable)IncludedTotal monthly cost~$80/month$500/month The biggest cost is your time. I spent a weekend and several evenings building the first version. However, I now have an asset that I can modify and reuse for any client without my costs increasing. The hidden costs are real: there’s no customer support, and you are responsible for maintenance. But for many, the savings and customization are worth it. Dig deeper: AI agents in SEO: A practical workflow walkthrough See the complete picture of your search visibility. Track, optimize, and win in Google and AI search from one platform. Start Free Trial Get started with Should you build your own tool? This approach isn’t for everyone. Here’s a simple guide to help you decide. Build your own if: You need a custom testing method that no SaaS tool offers. You want a white-labeled tool for your agency. Your budget is tight, but you have the time to invest in the process. Stick with a SaaS tool if: Your time is more valuable than the monthly subscription fee. You need enterprise-level security and dedicated support. Standard, off-the-shelf features are good enough for your needs. For many SEOs, the answer is clear. The ability to build a tool that works exactly the way you do, for less than $100 a month, is a game-changer. The process will be frustrating at times, but you will end up with something that gives you a unique advantage. The era of the practitioner-developer is here. It’s time to start building. View the full article
-
National Burrito Day freebees and deals: Where to get your Mexican food fix today, from Chipotle to Del Tacos
Burritos should be celebrated, because they are ingenious inventions that wrap deliciousness in a handy tortilla for extra convenience. National Burrito Day (today, Thursday, April 2, 2026) was created to do just that. Here’s a little history about the origins of the yumminess before we dive into the freebies and deals to observe this glorious unofficial holiday. A brief history of burritos Burritos hail from Northern Mexico and were invented in the early 20th century. The region’s climate was ideal for wheat, so larger tortillas were made out of the crop. This bigger vessel set the scene for a new culinary delight. The first burritos contained meat, beans, and cheese. The word “burrito” in Spanish literally means “little donkeys.” In Northern Mexico, donkeys were used to carry people and important cargo, much like the tortilla carries the filling of the burrito. It is believed that this is where the name comes from. The burrito was especially popular in California. America put its own stamp on the dish, adding ingredients such as sour cream, guacamole, rice, and more. In 1960s San Francisco, the “mission burrito” was popular among workers. Decades later, this same dish would inspire Steve Ells, founder of Chipotle Mexican Grill, to open the first store in Denver, Colorado, in 1993. These days, it’s fair to say the burrito is fully immersed in the cultural zeitgeist. List of National Burrito Day 2016 freebies and deals The best way to celebrate National Burrito Day is through your stomach, of course. Lots of restaurants have deals to help you do just that. Chipotle Mexican Grill rewards members have the opportunity to get free delivery on the big day via the restaurants’ website or app. Use the code DELIVER for the burritos to come to you. El Pollo Loco wants you to celebrate with friends. It has a buy one, get one free (BOGO) burrito deal. At Del Taco, reward members are eligible for a free classic burrito with a $3 purchase. Margaritas Mexican Restaurant rewards members will get 50% off of any burrito. This includes the El Jefe Burrito, which includes three different meats, and the Burrito Vegetariano, which has none. If you are craving a breakfast burrito, head to Torchy’s Tacos. It is offering a $5 option until 2 p.m. for customers who dine in. Red Sox Fans can even get in on the action. David “Big Papi” Ortiz has teamed up with Anna’s Taqueria to create the Big Papi Burrito, which will debut on the big day. This burrito gives back, as $1 from each burrito sold will be donated to the David Ortiz Children’s Fund. This organization helps provide “lifesaving heart surgeries and care to children in the Dominican Republic and New England.” Wherever you decide to celebrate, enjoy your burrito! View the full article
-
Visa says AI could start making purchases for you. Not everyone wants that, but here’s how close we are
What if you didn’t actually decide to buy that last thing in your cart? A report from Visa released on Thursday suggests that, in some cases, you might not have. According to a survey from the financial services company, artificial intelligence is no longer just helping people shop. In many cases, AI is starting to shape what people buy, and in some cases, even act on their behalf. The research is based on surveys of both U.S. consumers and business decision-makers. It shows that AI systems are moving from assistants to participants in commerce. That influence is already showing up in everyday behavior. Nearly 40% of Americans say they have made a purchase they would not have otherwise considered because of an AI tool. The impact tends to happen early in the shopping process, where AI surfaces options, compares products, and narrows choices before a person checks out. AI becomes a new kind of customer For Visa, this shift represents a new type of customer. “AI is not just technology. It’s a new customer segment. Your next cancelation or your your next review or your next booking of travel will increasingly be made by an agent, or at least with the help of an agent,” Visa CMO Frank Cooper III tells Fast Company. As that dynamic takes hold, companies are no longer just competing for human attention. They are also competing for machine selection. The report refers to this next phase as “B2AI,” where companies are selling not just to people but also to AI agents acting on their behalf. Businesses are already preparing More than half of business leaders surveyed say they would allow AI systems to negotiate prices or terms directly with other AI systems. Many are also willing to share data, such as pricing and inventory, to support those interactions. At the same time, most organizations say they are already using AI in some capacity or testing it across functions like marketing, e-commerce, and payments. The shift is also changing how companies think about persuasion. Traditional approaches that rely on emotion or limited attention don’t translate the same way when the “customer” is a machine. “Well, the machine has no emotions, right?,” says Cooper. Instead, companies are focusing more on making information clear, structured, and easy for algorithms to process. Consumers are on board, but cautious Consumers are open to using AI in shopping, but with limits. According to the report, most people are comfortable letting AI handle tasks like comparing prices or applying discounts. More than half say they are okay with those uses. That comfort level drops pretty dramatically when it comes to AI actually spending money. Only about a third are comfortable with AI completing a purchase, and fewer are willing to let it spend without approval. That gap points to a broader issue: trust. “Agentic commerce does not scale unless consumers cross that trust threshold,” says Cooper. While 53% of consumers are already using AI to help them shop, only 27% are comfortable with fully autonomous behavior. The hesitation often comes down to control, visibility, and risk. Trust is the sticking point Trust is likely to determine how quickly this type of commerce grows. “That trust is the scarce commodity in this new AI agentic commerce environment,” Cooper says. Similar patterns have played out in earlier shifts in financial technology, where adoption picked up once systems became more familiar and widely used. “And so we think something similar is happening here,” he says. Consumers say they want the ability to intervene, reverse transactions, and control spending. They also tend to trust AI systems tied to financial institutions or payment networks more than independent tools. Brand still plays a role in that environment. “Brand is always important because it’s shorthand for something meaningful to people,” says Cooper. That matters especially when people are unsure how systems work or what happens if something goes wrong. Humans still shape the experience The report also suggests AI is unlikely to fully replace human-driven shopping—at least not for a while. “I do not believe, and we’ve seen no evidence, that people, consumers, want exclusively an experience where AI agents are buying everything,” says Cooper. Instead, different types of purchases may be handled differently. Routine or low-stakes items may be delegated to AI more often, while more personal decisions remain hands-on. “The more it’s just a basic commodity with no emotional energy in it . . . the more the machine or the agent will buy it,” he says. So while you might use AI to buy your next bottle of laundry detergent, you might be less likely to have it buy your summer wardrobe without okaying the purchase. For now, businesses are building toward a future where AI plays a larger role in transactions, while consumers are gradually deciding how much control they are comfortable handing over. View the full article
-
Revolut fined in Italy over misleading fee information in adverts
Competition watchdog singles out advertisements suggesting customers could trade with no commission View the full article
-
6 Reasons Why Cloudflare’s EmDash Can’t Compete With WordPress via @sejournal, @martinibuster
Cloudflare's WordPress competitor EmDash may be the future of CMSs, but there are six reasons why it's not that CMS today. The post 6 Reasons Why Cloudflare’s EmDash Can’t Compete With WordPress appeared first on Search Engine Journal. View the full article
-
Finally, a wearable designed for women approaching menopause
Ali Hewson, like many women, was unprepared for the intensity of perimenopause. “I initially thought I would just grin and bear it, as my mother did,” she says. But the symptoms piled on: brain fog, hot flashes, mood swings, dryness, exhaustion. “One minute you are happy and content, then suddenly you are anxious and irritable followed by intense heat and sweating,” Hewson says. At their worst, her hot flashes were happening hourly. The symptoms began to take a toll on her rich and varied responsibilities as a humanitarian activist, fashion entrepreneur, mother of four, daughter to aging parents, and wife to world-famous rock star Bono. Even after deciding to seek relief, Hewson faced challenges in finding the right care. She visited four different gynecologists, each of whom prescribed a different version of hormone replacement therapy, or HRT. “It was a journey over 10 years to find the best doctor and formula that works for me,” she says. “I felt like an experiment.” It’s this very scenario that inspired the development of a new wearable for women in perimenopause, dubbed Peri, which Hewson has backed as an angel investor. (In 2023, I wrote about the booming business of menopause and the ways in which entrepreneurs were pursuing the $16 billion opportunity.) Launching stateside this week, Peri attaches to the torso and uses proprietary algorithms to monitor—in a first for a wearable—hot flashes and night sweats. The device, which is about the size of a flattened AirPods case, stays in place with a sticker and is designed to be worn all day, night, and even in the shower. In addition to perimenopause’s signature vasomotor symptoms, it also captures sleep, exercise, anxiety, and periods. In a companion app, women can log lifestyle variables, like caffeine and alcohol consumption, as well as supplement use and medication. Over time, Peri packages this data into insights designed to help women as they progress toward menopause itself, which is defined as 12 consecutive months without a period. “When I heard about Peri, it was a light bulb moment,” Hewson says. “Of course busy women can’t be expected to record every single anxiety attack, hot flash, or sleep disturbance accurately.” Instead of guesswork, she imagined that Peri would be able to provide individualized answers to the question of how hormones, lifestyle, and symptoms interact. That same vision is what brought together Ireland-based Peri cofounders Heidi Davis and Donal O’Gorman, experienced scientists with a shared interest in women’s health and longevity. “For us, it was all about getting real data for women to navigate [perimenopause] with more confidence, and to not waste their money on things that aren’t working for them,” says Davis, who is CEO. Demand for wearables continues to grow, thanks to consumers’ interest in optimizing their health and wellness. Whoop, the wristband wearable built for fitness, raised $575 million at a $10.1 billion valuation earlier this week. Last fall, Oura, which makes a ring wearable, raised $900 million at an $11 billion valuation. Oura, like some other wearables, offers some data related to women’s health, such as period and peak fertility prediction. Peri aims to capture a more mature consumer market: the estimated 1.3 million to 2 million women in the U.S. who enter perimenopause each year, with their symptoms lasting as long as a decade. Wearables like Whoop, endorsed by star athletes, promise performance maxxing; most women in the perimenopause demographic, in contrast, would be happy to just feel like themselves again. New research, old problem Once considered an afterthought, perimenopause is now believed by scientists to set the stage for risk factors later in life, with severe vasomotor symptoms like hot flashes correlated with cardiovascular and other diseases. That shift in thinking is shaping new research, but reciprocal changes to clinical practice have been slow, at best. “Women are quite intuitive, they know something is not right, but they’re not getting the help they need,” Davis says. She and O’Gorman hypothesized that personalized data could bridge the gap between women’s experiences and their clinicians’ recommendations. Thus far, personalization has played little role in the booming menopause market, which is expected to grow from $18 billion in 2024 to $27 billion in 2030, according to Straits Research. Much of that growth has come from menopause influencers peddling lotions, serums, lubes, and supplements. But there are signs that women are eager for solutions that integrate with providers, particularly as interest in HRT continues to rise. In the U.S., women’s telehealth startup Midi Health raised $100 million in Series D funding at a $1 billion valuation in February. Peri has been operating lean. To date, the startup has raised 3 million euros (roughly $3.5 million) from Hewson, other angels and strategic investors, and Ireland-based impact fund WakeUp Capital. Davis and O’Gorman started testing an early version of Peri six years ago. “We had done the science and believed that if we had sensors in the right location, we could build algorithms for hot flashes and night sweats,” Davis says. They approached hardware specialist Analog Devices, which sent them a sensor circuit. “It was big and clunky and really ugly,” Davis says, but the data was “really good.” A first prototype followed a month later, setting the stage for Davis and O’Gorman to partner with Ireland’s Health Innovation Hub, which helps startups run trials. A WEARABLE FOR WOMEN Peri contains four sensors: motion, optical, electrodermal, and temperature. The noisy data that they generate can be used to infer heart rate variability and a host of other data points, using a methodology now common to Oura rings, Apple Watches, and other devices. To meet women’s specific needs and offer value beyond established wearables, Davis and O’Gorman decided to first isolate night sweats. Women in Peri’s trial manually logged their night sweats while wearing the prototype, and then Peri built algorithms capable of identifying the symptom’s biosignature. Crucially, Peri needed to be able to distinguish between symptoms and activities like exercise or household chores. “People sweat, their temperature goes up, their heart rate might change: Was that exercise or not?” says O’Gorman, who serves as COO. Initially, O’Gorman and Davis envisioned that Peri would function as a medical device delivered to patients through clinicians. But the trials changed their plans. Among the women participating, “the frustration was palpable,” O’Gorman says. They wanted a direct understanding of their own symptoms and the chance to regain a sense of control. The data “has clinical utility,” O’Gorman says. “But it will actually have a more impactful utility if it is put in the hands of women who can then decide what they want to do with that information.” PUTTING PERI TO THE TEST Last month I tried Peri for myself. At 42, I’m starting to think about perimenopause and regularly hearing from friends who have started on HRT. Earlier this year I wondered if I was having a first hot flash. At first glance, Peri is larger than wearables like a ring or the face of a watch. But once on, it’s almost invisible. The device is incredibly light and thin enough to seamlessly fit under most clothes. At bedtime one night, it nudged against my ribs when I rolled over and I realized I had forgotten about it entirely. If my symptoms were to progress, I can imagine wearing it for long stretches of time, as Peri advises. It’s safe to wear the device while showering or bathing, and the sticker that holds it in place can last up to two weeks. Many wearables require a subscription, but Peri is opting for a different model. The device costs a onetime fee of $449 (it’s HSA and FSA eligible), and it comes with two rechargeable batteries, a charger, and enough stickers for several months of wear. Simple charts in the app summarize symptom incidents each day and over time. In particular, I liked a feature that would allow me to compare my symptom incidence to that of the Peri population. For example, a symptom like anxiety can be difficult to pin down and quantify. With Peri, data enters the picture: Is my anxiety low or high compared to my peers? Has it increased over time? Without that contextualized information, I might feel hesitant to seek help from a clinician. As Peri goes live in the U.S., with Ireland and the U.K. to follow later this year, Davis is already looking ahead. She would like to make the device even smaller, in keeping with her mantra of helping women live their lives as normal while Peri works in the background. She’s also eager to refine the tailored insights that Peri plans to provide about when and why symptoms might be changing. “We’re building a database that doesn’t exist, and this database is going to help women overall,” she says. View the full article
-
Farage sacks Reform UK housing spokesperson over ‘everyone dies’ Grenfell comments
Simon Dudley’s remarks provoked anger from bereaved familiesView the full article
-
Nick Candy sells Chelsea mansion for more than £275mn
Deal for Providence House in central London marks capital’s most expensive house saleView the full article
-
Why are designers, engineers, and product managers in a ‘three-way standoff’?
A newsletter about the state of the product job market recently went viral in the design corner of the internet. It’s exposing a widespread debate about whether the role of the designer is narrowing in the age of AI. On March 24, Lenny Rachitsky, a former Airbnb product developer and author of the business Substack Lenny’s Newsletter, published an article featuring exclusive data on the state of tech hiring in early 2026. The data was collected by TrueUp, a tech job marketplace tracker. Overall, it paints a positive picture for the tech job market. But for designers it points to a moment of hiring uncertainty. TrueUp found that design roles have plateaued since early 2023, and ever since then, demand for product managers (PMs), the professionals who help guide a product from ideation to completion, has risen. These findings have ignited a debate online about how AI might be fundamentally changing the organizational chart at tech companies—and whether it’s making designers obsolete. Everyone from tech CEOs to designers at AI companies and Marc Andreessen, the cofounder of one of the world’s largest venture capital firms, are weighing in. Here’s what you need to know about the data and the larger debate. Inside the data: Design roles are hitting a plateau TrueUp’s data is collected by tracking job openings at “the majority of tech companies and top startups,” which includes more than 9,000 companies (not including consultancies or non-tech companies). According to Rachitsky, who has analyzed that data for the past four years, 2026’s outlook is, “surprisingly, the most optimistic” so far. To start, open PM jobs are at the highest levels they’ve reached since 2022: around 7,300 roles globally. Software engineer jobs are also trending up since a recent low in 2023, with 67,000 jobs available globally and 26,000 in the U.S. alone. “We don’t know if there would have been more open roles if not for AI or if AI is actually leading to more open roles, but since the start of this year, the increase in open eng roles is accelerating even more,” Rachitsky’s newsletter reads. And “AI jobs,” which include open roles at AI-driven companies as well as AI-specific roles at non-AI companies, are skyrocketing. There are currently 36,686 open AI jobs, compared to sub-10,000 numbers in early 2023. Amidst this general upturn, design jobs are having a less optimistic moment. Unlike PM and engineering, Rachitsky’s analysis notes that open design jobs have been relatively flat since early 2023. At the time of the newsletter’s release, TrueUp found just 5,700 roles available globally. From a macro perspective, the ratio of demand for PMs versus designers has flipped: In mid-2023, open PM roles overtook open designer roles, and the disparity has been increasing ever since. “I don’t know exactly what’s going on here, but it does feel AI-related,” Rachitsky writes in his newsletter. “Unlike PM and eng, which started growing in 2024 (two years post-ChatGPT), design didn’t. If I had to venture a theory, I’d say that because AI is allowing engineers to move so quickly, there’s less opportunity—and less desire—to involve the traditional design process.” How the design community is responding In the week following Rachitsky’s post on X, his analysis has been reposted hundreds of times and attracted an influx of discourse from the online community of designers, PMs, and engineers. Some responders see this industry data as a signal that AI is fundamentally changing designers’ workflows, and those who fail to adapt to the times are getting left behind. “Designers have designed themselves out of the equation because of design systems,” Roger Wong, head of design at BuildOps, commented under Rachitsky’s original post. “But, IMHO, the secret sauce has never been the UI. It was the workflows and looking across the experience holistically.” Claire Vo, founder of the AI copilot ChatPRD, added, “Often design teams & designers are the most resistant to change org in the EPD triad, with highly vocal AI opponents, and little skill or interest in the art of campaigning for influence or resources.” Most teams, she continued, treat design “like a tax they don’t want to pay.” “If a PM or engineer can get 85% there with tailwind and a dream, you better come to the table with more than ‘I represent the user,’” she concluded. Others believe that, in the long run, a greater reliance on AI tools will make human designers more important as tastemakers. “Design seems to be viewed as dispensable in this very moment,” wrote Jordan Singer, CEO of the computing company Mainframe. “But the reality that will become clear, as time has shown, is that design is what will make you rise above the rest.” On March 30, Rachitsky followed up on his newsletter through an interview with Andreessen, cofounder of Andreessen Horowitz. Andreessen likened the current atmosphere to a three-way standoff among product engineers, designers, and coders: each side of the triad believes that AI has given them enough tools to subsume the roles of the other two. “What’s so interesting about this Mexican stand-off is that they’re all kind of correct,” Andreessen said. “AI is actually now a really good coder, a really good designer, and a really good product manager.” While the future of this stand-off remains murky, it seems clear that the industry is currently in the middle of an organizational flux. New AI tools are constantly blurring the lines between these three roles, creating new positions that blend elements of each. In the future, there will almost definitely still be need for human coding, designing, and product managing skills—but we may not define each of those jobs the same way. In a recent interview for the Fast Company podcast By Design, Anthropic’s chief design officer, Joel Lewenstein, summed up this shift: “I think there’s a lot of role collapse at the very beginning, but there are still pretty clear swim lanes as things get into the later stages of product development.” PMs are still the best at figuring out a product’s business case; engineers are still the best at deploying those products; and designers are still the best at tackling bigger human-computer interaction questions, he said. “It’s like a Venn diagram that’s coming closer together.” View the full article
-
An all too special relationship
Maga leaders obsess over an imagined version of a long-lost BritainView the full article
-
Double-pledging risk: What mortgage lenders should know
Recent double-pledging scandals in auto lending and the U.K. put U.S. mortgage lenders on alert. Here's what to watch and how MERS, e-notes and electronic vaults can help. View the full article
-
How Disney Imagineers are using AI and robotics to reshape the company’s theme parks
With last weekend’s opening of World of Frozen in the renamed Disney Adventure World park, Paris became the new leader in advanced technology among the company’s theme parks. It’s a title that shifts hands frequently, but with its robotic Olaf and a new nighttime show that blends airborne and water drones with fountains, fire, and water walls, Adventure World is a technological marvel. Disney tends to downplay the focus on technology in its park attractions. Workers and executives at Walt Disney Imagineering (WDI) see the technology as a way to evoke emotion, their primary goal. And with a growing arsenal of tools at their disposal, from AI to powerful game engines (along with plenty of homegrown methods), all of the parks have a lot of ways to summon feelings from guests. Disneyland Paris has an additional advantage up its sleeve: It sits just a few hours from Disney’s R&D hub in Zurich, where much of the company’s robotics work is happening. That proximity helped shape the new Olaf. The project began with StellaLou, a theme-park-exclusive character popular in Asia, whose ballerina persona led the team to build a robot capable of a full pirouette. But Imagineers decided she wouldn’t resonate as broadly as Olaf, so they shifted the work onto a more widely recognized character and kept pushing the technology forward. Homecoming The Olaf robot shares the same distinctive walk and mannerisms as he does in the Frozen films. He’s the latest advance from the division that was the birthplace of the BDX droids that debuted two years ago (and now are appearing at most of the company’s parks) and the real-world Herbie robot inspired by Fantastic Four. He represents more than a shift from the audio animatronics the company is famous for. He’s also emblematic of a new type of thinking at Disney Imagineering. That shift starts at the top. Bruce Vaughn was named president and chief creative officer of Walt Disney Imagineering three years ago. Before that, he spent 22 years as an Imagineer, taking a seven-year break between stints with the company to explore the entrepreneurial world. Once he was lured back to Disney, he says, he brought the startup culture with him. “We’ve actually culturally shifted Imagineering by . . . celebrating anybody who finds opportunities and creates opportunities,” he tells Fast Company. “Just because we’ve spent 74 years . . . doing things one way, that doesn’t mean it’s the best way, especially given [the] new tools.” And make no mistake, Disney is leaning into those tools heavily. AI and Imagineering AI critics might cringe at the thought of the technology being incorporated into the very heart of Disney’s experiential creativity unit. Vaughn says it has served as a tremendous asset for his team, but only as a supplement to the team’s inherent creativity. “We did a side-by-side test internally to see if AI could generate anything that we would build,” he says. “It sparked ideas. Then, when we went one step further and we put it in the hands of someone who can actually sketch and draw, they at first were like, ‘I don’t know if this would be useful.’ Very quickly, they were like, ‘Oh my god, what I can do now in days used to take me months.’” The tech industry is currently facing supply chain issues with RAM and tremendous demand for GPU chips, which is threatening to slow the progress of some AI companies. Vaughn, though, says Disney’s unique position in the entertainment and technology worlds minimizes those industry headaches. “We have a very tight relationship with Jensen [Huang, CEO of Nvidia], quite frankly,” Vaughn says. “I think it’s partly because of the competitive advantage that we have at Disney and Imagineering.” Reinforcement Learning For Olaf and other robotics, AI has been especially useful. To ensure Olaf doesn’t fall flat on his snowy face while ambling around, Imagineers relied on reinforcement learning, an AI-driven tool that lets robotic units make optimal decisions through trial and error. It’s a way for AI to learn without having to constantly tweak programming—and without it, Disney’s robotics unit wouldn’t be anywhere close to where it is now. Over the long term, says Kyle Laughlin, SVP of Walt Disney Imagineering R&D, the advances you’re seeing with robotics will be incorporated into newer animatronics on rides. “Our robotic characters and our animatronic characters are going to begin to merge,” he says. Most animatronics are bolted to the ground, but there’s a growing interest at Disney in expanding their range of motion. For example, an animatronic of founder Walt Disney unveiled at Disneyland in California last year takes a few steps onstage in his show. And there are plenty of Star Wars droids that would be hard to re-create in a character costume. The future of Olaf Olaf, for now, appears only at World of Frozen in Paris and Hong Kong. And visitors might be upset to learn that at present, they’ll primarily see him at a show in the lagoon of both parks. Kids who want to give him a warm hug won’t get the opportunity. That’s not expected to be permanent, though. “He’s so popular that we have to ensure, both from a security perspective as well as an operational one, that we can ease him into the park,” Laughlin says. “You’ll absolutely see him roaming the park in the future.” Will there be meet and greets with Olaf? “Yes. I mean, absolutely yes,” he says. “The North Star goal that we have is to be able to have that huggable moment.” Olaf won’t remain geographically inconvenient for Americans, either. Just as it did with its BDX droids, which started at one park and spread, Imagineering plans to roll out Olaf robots to parks and cruise ships globally, though there’s no timeline for that just yet. “He’s one of our most popular characters, so domestically you’ll also see him as well,” says Laughlin. “That’s really kind of the important point: We really are building these now for operations, and to ensure that they’re everywhere.” Expect more robots as well. A Lion King area is currently under construction at Disney Adventure World—and putting people in a lion costume could be underwhelming. Laughlin also mused about the possibility of a robotic Sven, Kristof’s reindeer companion in Frozen. From their early days in the Tiki Room and the Carousel of Progress, robotics have been a critical part of Disney’s competitive advantage in the theme park space. With the rollout of these new robotic characters, the company is looking to stay at the head of the pack. “All of our innovation [follows] a planned, critical path where it goes into the product,” Vaughn says. “It isn’t just, ‘Hey, we’re an experiment. Maybe we’ll end up doing something.’ We literally commit to an early idea. And since we can move much more quickly now, based on our relationships with companies like Nvidia and Unreal Engine, we can move an order of magnitude faster.” View the full article
-
Tostitos redesigned its bags to emphasize one obvious thing
Last year, PepsiCo started printing real potatoes onto every bag of Lay’s. The reason? In a world where people are increasingly concerned about the provenance of their food, 42% of the population didn’t realize that the world’s most popular potato chip was made from potatoes. So they put a potato on the packaging. And now, the company is updating Tostitos bags—the most popular plain tortilla chip in the world—with a similar strategy. While Lay’s got a dose of potatoes, naturally, all Tostitos bags feature corn. “We started by being really honest with ourselves. The research was telling us that the old packaging wasn’t working—it was actually reinforcing a lot of the wrong perceptions,” writes Hernán Tantardini, CMO of PepsiCo Foods, over email. “People saw Tostitos as a party brand. The quality and craft story wasn’t coming through at all.” Technically speaking, Tostitos are classified as an ultra processed food, due to their use of refined seed oils. But they still feature a stupid simple and clear ingredient list: Corn, oil, and salt. 71% of consumers are reading labels more closely than before, and the front of the bag is a gateway to the back. The bags used to read “no artificial flavors, colors, or preservatives” right up top. Now they promise “Masa made in the traditional way,” with these other notes sidelined. The result is repositioning Tostitos as a more authentic and culturally-born product, anchoring Tostitos in the old way of doing things—which aside from signaling quality and cachet, tends to be more natural. As for the actual corn, that’s presented in an entirely different way than Lay’s. For Lay’s, PepsiCo photographed countless real potatoes in different presentations. For Tostitos, it opted for illustration—to tie it back to that idea of masa production, a complex process known as nixtamalization in which corn is treated with lime to make it more digestible and nutritious for consumption in tortillas, sopes, and other Mexican delicacies. “We looked at photography, but the more we explored it, the more it felt like it belonged to a different brand,” says Tantardini. “Tostitos has this warmth to it—this sense of joy and togetherness that’s been a part of its DNA forever. Photography felt too polished, too literal. It would’ve flattened something that’s actually quite live.” The illustration has an imperfect, hand painted feel—and your eye reads a human touch beneath the perfectly machined Tostitos wordmark. I find myself wishing that PepsiCo went even further, and embraced xilografía (the woodblock printing out of Mexico) with elements like the window frame or even subheading labels. But the two tone kernels and cobs of corn on Scoops and Street Corn varieties really are quite pretty for a mass market snack chip. While many of the colors are technically the same on the old and new packages, the chosen hues are softer and intentionally read warmer, with a basis in earth tones, according to the company. That said, Tostitos doesn’t read like some half-apologetic Trader Joe’s snack brand, unsure if it’s there to party or to apologize for the indulgence. The bags still read like a celebration. In a final, playful twist, the front window—which reveals the actual chips through the bag—now dips itself right into a large bowl of salsa or guacamole sitting below. (The old version featured photorealistic jarred Tostitos salsa on the side, like an overzealous advertisement.) This entire approach to craft could help Tostitos—which has ceded a few percentage points in sales volume since raising prices in 2022/23—compete with smaller batch brands that have cut ever so slightly into its market share. PepsiCo did validate the approach in test stores, while consumers will see the new rollout over the coming months. “From a business perspective, this is really about changing perception where it matters most,” says Tantardini. “If we can use packaging to clearly signal craft, quality, and care, we can rebuild confidence in the brand and lower the barrier for people who’ve drifted away because they didn’t realize the craft and care that goes into our chips and dips. That’s the opportunity. And that’s what we designed for.” View the full article
-
The unexpected childhood activity that predicted my career path
“And a cascade of lace here, here, and here.” I thwacked my pen against the notepad to emphasize each word while my cousin nodded vigorously. At 8 and 10, we carefully reviewed our wedding dress designs as if our big days were just moments away. While our parents prepped dinner, we rehearsed our grand bridal entry in painstaking detail. I’m probably not the only person who had this fantasy when I was little, but what I didn’t realize was just how that role-play would translate into the career that I have right now. It all started with my own elopement in 2021, and the subsequent blow-out bash a year later. My husband and I juggled countless chaotic spreadsheets, email chains, invoices, journals, and Post-its. I felt overwhelmed by the lack of tools to orchestrate this complex logistical feat, and I realized that I could make a living by fixing this chaos. My unconventional career path What I loved about weddings were the systems and the operational aspect. But in addition to planning a wedding, it took me a few jobs to understand that. The summer after I graduated from college, I worked as a tutor for middle and high school kids while I looked for a full-time job. Just days in, it became clear that the tutoring company was in full operational shambles. I started mapping out a plan to transition the company to electronic records, using homework assignments and practice flashcards. My next job was at a boutique marketing firm. A few months on the job, I realized it was on the scrappier end of the spectrum than I initially expected. Soon, the quirky operational novelties I had loved at first were starting to make me itch. I started to scheme to make things more efficient again. My next role allowed my skills to find a true home. I was leading a team of operations specialists at the unicorn startup, Carta. This experience would end up being crucial to my wedding planning business that I’d later start. Doing my best work in chaos The refrain that continued in my mind over the many months of wedding planning was, ‘Why is this so hard?’ When about 2.5 million people get married in the United States every year, it felt like I (and every engaged person I knew) had to reinvent the wheel while undergoing a deeply frustrating and time-consuming exercise. And as my husband and I lived through the organizational chaos of our wedding planning journey, I began to think to myself, ‘someone should really fix this problem. Someone who is obsessed with the details, thinks like an operator and a consumer, and someone who loves bringing clarity to chaos.’ That was the moment I began to connect the dots together. Without any context, someone can look at my professional history and feel disconnected. But at that moment, I’d realized that all my personal and professional experiences up to this point had led me to this challenge, and that I was ready to meet it head-on. Your interests can provide career clues Before I knew what careers existed, the common threads of how I thought and how I solved problems were there right in front of me. I took the concept of loving systems and organization as a personal interest rather than a professional secret weapon. At times, I interpreted my drive to improve operational circumstances as a distraction to the job at hand. There were also instances where it felt like an inconvenience to my managers (who weren’t always receptive to it). My previous roles didn’t grant me authority or purview over the messy systems or inefficient processes, but I couldn’t stop engaging with it. I didn’t realize how unique my perspective was. Identifying inefficiency came naturally and easily to me, so I naturally assumed that it was the same for other people. Why it’s important to pay attention to what you think When deciding how to approach the professional world, people ask us, ‘What are you passionate about?’ or ‘What do you want to be?’ We build resumes and write cover letters, highlighting the projects we led and the metrics we moved. But in the process, we overlook what feels natural. We’ve placed ourselves in organizations and teams and roles, but never stopped to ask ourselves the critical question: How do I think? What challenges are my brain hard-wired to solve? What is the recurring thread in my life that I haven’t paused long enough to see? The throughline is usually there. We just don’t think to look for it. I close my crisp notebook with a satisfying snap. I’ve just finished explaining to my cousin that the processional music must begin exactly as Mr. Bear arrives at the steps, so that it will reach the perfect crescendo as he reaches the altar. The chairs are straight. The timing is perfect. The experience unfolds exactly as planned. In hindsight, I wasn’t pretending to be a bride. I was designing flow, sequencing emotion, and building structure around what is supposed to be a joyful and effortless moment. Long before I had the language for systems, operations, or user experience, I was already drawn to the architecture beneath the celebration. Maybe you had your version of playing wedding as a child, and it looks completely different from mine. Look closely and think about what it is about that scenario that interests you. Because if you look closely, you might just have been rehearsing what you’re meant to do all along. View the full article
-
How NASA designed the Artemis II space suits for a worst-case scenario
“Houston, we have a problem.” The misquoted phrase is so ingrained in popular culture that it has become the standard comeback to any unexpected mishap. It’s also the last phrase NASA’s Artemis II mission control wants to hear in the coming days because, unlike those of us on Earthly terrain, an astronaut midway to the moon won’t be muttering it after they accidentally burn their toast. A four-person crew took off from Kennedy Space Center in Florida on April 1 for NASA’s first lunar flyby since Apollo 17 in 1972. The organization has done everything it can to ensure the safety of the astronauts, knowing that any harm to the courageous humans could set its lunar program back many years, or cancel it altogether. One part of its insurance policy is a new space suit that’s designed to sustain the Artemis II crew for six days—enough time to go to the moon and back—in case there’s a catastrophic event in their Orion spacecraft. Jeremy HansenChristina KochVictor GloverReid Wiseman A lifeboat in a space suit When Jack Swigert, command module pilot of Apollo 13, radioed “Houston, we’ve had a problem here” on April 13, 1970, an oxygen tank explosion had just severely damaged the spacecraft just 56 hours into its journey to the moon. The astronauts on board couldn’t simply pull a U-turn 200,000 miles away from Earth. And since they didn’t have enough oxygen, Swigert, along with commander Jim Lovell and lunar module pilot Fred Haise, abandoned their crippled spaceship and hunkered down inside the lunar lander, using it as a makeshift lifeboat for the harrowing trip home. But the Artemis II mission—a roughly 10-day loop around the moon—flies without a lunar lander. If the Orion capsule’s hull breaches for any reason and vents its breathable air into the void, the crew has nowhere else to go. NASA’s answer was to build a lifeboat of sorts directly into their suits. For this return to the moon, the space agency assumed such a leak could happen and they needed a last line of defense to keep the crew alive in a vacuum for a week. The suit gives astronauts a 144-hour survival window, the exact time required to abort a translunar flight, whip around the dark side, and coast back home. How it’s made The aptly named Orion Crew Survival System (OCSS) serves as this wearable sanctuary. According to the agency’s Crew Systems branch, “the suits can keep astronauts alive for up to six days if Orion were to lose cabin pressure during its journey, with interfaces that supply air and remove carbon dioxide.” Dustin Gohmert Dustin Gohmert, a mechanical engineer who worked on Space Shuttle garments before taking over the OCSS program at the Johnson Space Center, notes that the gear operates as an independent vehicle. “They become your own personal-sized spacecraft that can last up to six days,” he told CBS News. Victor Glover A NASA lab tour details the concept: “In effect, the space suit is a body-shaped balloon that holds your personal atmosphere,” strictly for use inside the ship. The suit plugs directly into the Orion capsule’s Environmental Control and Life Support System (ECLSS) via a thick umbilical cord. This artificial artery keeps the astronaut from overheating by pumping chilled water through an undergarment. Simultaneously, the capsule acts as a mechanical lung that regulates humidity, scrubs out deadly carbon dioxide, and forces a breathable nitrogen and oxygen mix into the helmet. Christina Koch Astronauts eat and take medicine inside the sealed balloon using a pass-through port built into their rigid helmet dome. They snap pouches of liquid food and water right into this valve. If someone falls ill during the six-day ordeal, the ship’s medical kit includes a specific tool that shoves pills through the same helmet port without venting the suit’s precious pressure. Each suit is meticulously tailored to the individual wearer and paired with custom-molded, shock-absorbing seats for launch and reentry. The suit also features a pleated fabric design hidden in the shoulders that unfolds when pressurized, giving the arms enough clearance to move. The gloves are spun from rugged materials that interact flawlessly with Orion’s digital touchscreens, while internal microphones and speakers are embedded directly into the helmet so the crew can communicate. To prevent snagging in the tight cabin, the communication wires run down a protected channel on the right leg. Floating in zero gravity inside a cramped cockpit means a loose cord can easily snag a critical flight switch and trigger a disaster. To prevent this, designers built an asymmetrical storage system directly into the fabric instead of using cargo pockets. The right thigh features a custom compartment that swallows the temperature-control dial and the thick tubes pumping ice water to the undergarment, locking them flush against the astronaut’s leg. Meanwhile, hidden channels route the electronic brain that controls the suit and the plumbing for human waste safely out of the way. What if the ship dies? There is only one problem with the suits, and there’s no way around fixing this one: They depend on the spaceship’s ECLSS. If the life support system fails, the astronauts will not survive no matter how well-designed their space suits are. To avoid that extreme scenario, the Orion capsule prevents a catastrophic ECLSS failure by implementing overlapping safety nets. Lockheed Martin—the designers and manufacturers of the ship—created a life support architecture with duplicate secondary pumps and backup valves that automatically kick in if the primary hardware chokes. The ship’s digital brains also have redundancies: Four identical flight computers run the show simultaneously. If a software glitch wipes out all four, a completely isolated fifth computer (running entirely different code) takes the wheel. If every single spacecraft system fails and the umbilical stops flowing, the astronaut relies on something called the bailout bottle. This suit-integrated emergency oxygen tank holds a tiny amount of breathable air—just enough to do one last Hail Mary operation, like switching to a different ECLSS oxygen line or getting out of the capsule after crashing into the ocean. During an emergency ocean splashdown, the OCSS transforms into a heavy-duty maritime survival rig. It inherits its blazing pumpkin-orange hue from the old Space Shuttle Advanced Crew Escape Suit (ACES)—a color specifically chosen so pilots in rescue helicopters can easily spot a human bobbing in the open water. But where the old ACES gear only provided roughly 10 minutes of bailout air for low-Earth orbit emergencies, the OCSS packs an automatically inflating personal flotation device right into the architecture. Tethered to this life preserver is a meticulously packed emergency kit containing a personal locator beacon to ping rescue forces; a specialized rescue knife; and a comprehensive signaling stash equipped with a mirror, strobe light, flashlight, whistle, and chemical light sticks. But if the ECLSS collapses on the journey around the moon . . . that’s the end of the line for the crew. Which is why, when we’re reading or watching a report on how Artemis II is going, we need to pay close attention and think about the very real risks these four heroes are assuming, from the moment they strap themselves to a flying bomb full of 5.75 million pounds of explosive fuel all the way to the moment they blaze through the atmosphere and splash into the Pacific Ocean. View the full article
-
Artemis II: Why our return to the moon took so long
While I was leading a tour of the National Air and Space Museum in January 2026, a visitor posed this insightful question: “Why has it taken so long to return to the moon?” After all, NASA had the know-how and technology to send humans to the lunar surface more than 50 years ago as part of the Apollo program. And, as another tour guest reminded us, computers today can do so much more than they could back then, as evidenced by the smartphones most of us carry in our pockets. Shouldn’t it be easier to get to the moon than ever before? The truth is that sending humans into space safely continues to be difficult, especially as missions increase in complexity. New technologies require years of study, development, and testing before they can be certified for flight. And even then, systems and materials can behave in ways that surprise and worry engineers and mission planners; look no further than Boeing’s Starliner CFT mission or the performance of the Orion heat shield on Artemis I. Issues with Starliner’s thrusters led NASA to return the spacecraft from the International Space Station without its crew. Unanticipated chipping of the Orion heat shield resulted in years of research, culminating in NASA altering the atmospheric reentry plans for the Artemis II mission. NASA’s programs also require sustained political will and financial support across multiple presidential administrations, Congresses, and fiscal years. As a historian of human spaceflight, I have studied the space agency’s efforts to engage the broader public to convince American taxpayers that their programs hold value for the nation. NASA just launched its first crewed flight to the moon since the Apollo era: Artemis II. A crew of four will conduct a lunar flyby, laying the groundwork, the agency hopes, for a landing on the Artemis IV mission. The story of NASA’s effort to return humans to the moon is long and winding, demonstrating the complexities of turning grand ambitions into real missions. Post-Apollo In early 1970, with two successful moon landings on the books, President Richard Nixon sought to reduce NASA’s budget to better align with his administration’s priorities. This decision put the space agency in a difficult position, which ultimately led to the cancellation of three planned Apollo missions to conserve funding for its plans for long-term human activity in low-Earth orbit. NASA repurposed the third stage of a Saturn V rocket to create the first U.S. space station, Skylab, which operated from 1973 to 1974. The space agency used leftover Saturn IB rockets and Apollo command and service modules to send crews to the station. Over the next three decades, NASA developed and operated the space shuttle. The fleet of space shuttle orbiters supported satellite deployment and microgravity research on orbital missions of up to 17 days. This work was meant to enable future long-duration human missions and provide benefits to people on Earth. For example, data from protein crystal growth experiments have informed the development of medicines. The space shuttle program facilitated the construction, maintenance, and staffing of a continuously inhabited research platform in orbit, the International Space Station. The first modules launched in late 1998. Where to next? As the new millennium approached, the Clinton administration tasked NASA to think beyond the space station. What could robots and humans do next in space? And where could they do it? Notably, the White House expressed an interest in locations beyond low-Earth orbit. NASA, it turned out, was well-positioned to meet the administration’s request. Daniel Goldin, NASA administrator, was already thinking about preparing proposals for the next presidential administration and had recently sponsored a human lunar return study. In 1999, he established a team to investigate new technologies, missions and destinations for the 21st century. This work took on new significance following the tragic loss of the space shuttle Columbia crew in February 2003. Many people, including those in the new George W. Bush White House, wondered whether the human spaceflight program should continue—and if so, how. Administration discussions culminated in Bush’s Vision for Space Exploration in 2004, which directed NASA to retire the space shuttle after the completion of the space station. It called for returning humans to the moon on a crew exploration vehicle designed for destinations beyond low-Earth orbit. It also called for continuing robotic exploration of Mars and engaging companies and international partners in space. Fifteen years earlier, President George H. W. Bush had also announced a moon and Mars exploration program, but congressional concerns about cost kept space travelers close to home. The Constellation program’s legacy In December 2004, NASA began the process of finding a manufacturer for the crew exploration vehicle. By August 2006, the space agency awarded Lockheed Martin the contract to build the capsule, which it had named Orion—the same Orion planned to carry Artemis astronauts to the moon. Years of research, development, and testing followed for Orion as well as the Ares I crew and Ares V cargo launch vehicles. Together, these technologies made up the Constellation program. Constellation had two primary objectives: in the near term, to help transport crew to and from the space station after the space shuttle program ended; in the long term, to enable human lunar exploration. Building systems that could work in both Earth orbit and around the moon was supposed to save the time and cost of developing two vehicles. Similarly, adapting space shuttle program hardware could supposedly cut costs. During the first months of Barack Obama’s presidency in 2009, the administration initiated an independent review of NASA’s human spaceflight plans. The Augustine Committee, chaired by retired aerospace executive Norman Augustine, found that the agency’s ambitions outstripped its limited budget, leading to significant delays. The first Orion spacecraft was likely to arrive after the space station ceased operations. The committee proposed several paths forward at the current funding level, which prioritized space shuttle and space station programs. An additional annual investment of $3 billion would allow for human exploration beyond low-Earth orbit. Ultimately, the Obama administration canceled Constellation, but two of its technologies lived on, thanks to U.S. senators from states that would have been affected by cuts. The NASA Authorization Act of 2010 funded Orion’s continued development, shifting responsibility for space station crew transportation to commercial vehicles. It also directed NASA to develop the space launch system, a redesigned Ares V heavy booster, to send Orion to the moon. The technical strategy had political benefits, too, preserving jobs in numerous congressional districts by providing continuity for aerospace contractors. In December 2014, a Delta IV heavy rocket launched the first Orion capsule on a test flight, providing engineers with data on spacecraft systems and the heat shield. By October 2015, the space launch system had completed a critical design review, the last step before manufacturing could begin. Introducing Artemis In December 2017, the new The President administration issued a policy directive shifting the focus of NASA’s human spaceflight program back to the moon. The space agency would use Orion and the space launch system in a race to meet an ambitious 2024 landing date. NASA officially named the program Artemis in May 2019. The 25-day Artemis I mission, launched in November 2022, was a major milestone for the program. This uncrewed flight was the first flight of the space launch system and the first to integrate SLS and Orion. It laid the groundwork for Artemis II, the first crewed flight of the SLS. Over more than 50 years, each new presidential administration has reassessed the place of spaceflight among its priorities, either encouraging or curtailing NASA’s efforts to return humans to the lunar surface. Each crewed flight requires the alignment of technical expertise, political will, and financial support over years if not decades. For the space fans who watched the Artemis II launch, the wait for countdown may have felt long. But was just a blink in NASA’s long journey back to the moon. Emily A. Margolis is a curator of contemporary spaceflight at the National Air and Space Museum at the Smithsonian Institution. This article is republished from The Conversation under a Creative Commons license. Read the original article. View the full article
-
How AI agents are changing journalism
I’ve been using Claude Cowork extensively over the past month and a half. And not coincidentally, I’ve been more productive than I ever have in that same period. The shift to working agentically is something so profound, you really can’t understand it until you experience it for yourself. Just one example: As the operator of a business selling AI training courses online, email marketing is an important component of getting the word out about them. But much of the work is rote: segmenting my email list, creating templates, writing largely similar drafts, and scheduling them in my email provider—a piece of software I look forward to using about as much as a visit to the dentist. Now I hardly ever touch that software; Claude Cowork does it for me. When you have access to agents, you can loop them in on any computer task with three beautiful words: “You do it.” Now AI doesn’t just draft emails for me—it puts them in the campaign builder, targets the right audience, gets all the settings right, and then taps me on the shoulder (via a notification) so I can approve the work before it schedules the emails to go out. Once you start working with agents, you quickly start crossing things off your to-do list faster than ever before. From doing to directing This is not just faster work. It is different work. It shifts the focus from slogging through individual tasks to focusing on outcomes, assigning the actual execution to an army of digital workers, then reviewing what they’ve done. You essentially become the CEO of your job. So what happens to a newsroom when everyone starts working agentically? Over the past 30 years, reporters and editors have needed to become skilled at many different systems: project-management software for tracking stories, content management systems for publishing them, SEO plug-ins, social media management platforms—the list goes on. Working alongside agents, journalists can theoretically assign agents to deal with all of that while they go and do the important, human-centered work of reporting and editing. Where this gets uncomfortable is when this paradigm is applied to the writing itself. This came to a head recently with the uproar over what The Plain Dealer, Cleveland’s primary newspaper, is doing: leveraging an AI writing agent so reporters can simply feed notes and context to create stories. To be clear, all the stories are then edited, and the reporter has final say over the copy. But applying agents this way brings up hard questions about jobs, skill-building, and career paths. But even putting aside that specific use case, it seems inevitable that agents will eventually take on much of the production and distribution work around content and storytelling. Whether it’s social media management, SEO (and GEO), or getting all the little drop-down menus, boxes, and tag fields in your CMS just right—those are all jobs for agents. More importantly, roles that are centered around optimizing those tasks will gradually go away. If you think about it, that inherently devalues certain kinds of content. When search and social platforms drove audiences, newsrooms set up workflows around those patterns. Many roles emerged that were simply writing to a trend, publishing undifferentiated “quick hits” around trending topics to maximize clicks. Those roles were effectively hyper-optimizing production of formulaic stories, writing for algorithms and chasing virality through pattern recognition. That has very little value in a world where a robot can do all of that much faster than a human ever could. And this is the mechanism by which AI can actually be healthy for journalism, something I predicted in my first column. Agents are a crucible for knowledge work, burning away anything and everything that can be automated, leaving only the parts of the job that can’t be easily repeated—the work that requires either creating new information or judgment, context, and taste. The new shape of the newsroom If you were building an AI-first newsroom today with this idea at its core, virtually all roles would be centered on the parts of the job that are exclusively human: building trust with sources through access and relationships, doing original reporting and finding information that’s exclusive to your brand, determining what stories matter most to your audiences and what angles to take, and applying the art of storytelling to all of it. While that sounds idyllic in some ways, the reality is that with agents handling most of the execution, there will probably be less work to go around. In almost all cases, organizations will be smaller, with different career paths, even if the work is richer. A constraint, for now, is access. Tools like Claude Cowork and Claude Code become truly powerful only when they can move beyond drafting and into systems (email, CMS, analytics, internal documents). That is where most organizations get uneasy. Granting an agent permission to act inside those environments raises questions about security and accountability. Most teams are still feeling their way through this, limiting agents to narrow tasks or read-only access. But that tension is temporary. As guardrails improve and familiarity grows, those permissions will expand, and with them, the scope of what agents can do. Once that happens, journalism does not lose its purpose. It comes into sharper focus. An AI-first newsroom doesn’t mean a less human one. In fact, it means the opposite. When the repeatable work is handled by machines, what remains is the work that defines the craft: earning trust, finding new information, and making sense of it for an audience. The uncomfortable part is that there may be fewer people doing that work. The hopeful part is that the work itself becomes more meaningful. View the full article
-
KPMG cleared by UK watchdog over audit of gambling group Entain
Decision is boost for Big Four firm seeking to repair reputation after series of scandalsView the full article
-
AI isn’t just reshaping productivity and threatening to kill jobs. It’s changing how we lead, communicate, and treat each other. It’s also creating a new gender gap
For nearly four years now, the conversation about generative AI has revolved almost exclusively around productivity, threatened jobs, automatable tasks, efficiency, and competitiveness. But there is a largely underestimated dimension to this revolution: its cultural effects. AI is not just transforming how we work; it is transforming how we are together, how we trust each other, how we communicate, and how we organize ourselves. To measure this, it helps to borrow a framework from Erin Meyer, a professor at INSEAD whose book The Culture Map identifies eight dimensions along which the cultures of the world differ. Applied to artificial intelligence, Meyer’s eight dimensions reveal a series of cultural shifts that are more profound than we know. 1. How We Communicate: AI Is Training Us to Say What We Mean Generative AI demands clarity. An effective prompt is an explicit one. There’s no room for body language. This constraint is gradually reshaping how we communicate with each other, too. Cultures that have traditionally relied on what is left unsaid—where reading between the lines or sensing the mood in the room is a valued skill—are being pushed toward greater explicitness. As AI mediates more exchanges, the richness of implicit communication erodes. And there is the curious rehabilitation of the typo. For decades, a spelling mistake in a professional message was a sign of carelessness, even disrespect. Not anymore. A typo is increasingly read as proof that you wrote it yourself—that you took the time, that you cared enough to type it out without outsourcing the task. Imperfection has become a signal of authenticity. 2. How We Give Feedback: AI as a Cultural Mediator and Sugar-Coater Large language models are not built to be brutal. They begin by finding something to praise, soften their critiques, and close on a constructive note. After thousands of interactions with tools that say “great question” before correcting your mistake, even cultures accustomed to blunt, direct feedback begin absorbing a more diplomatic register. But AI also has a more positive effect on collective evaluation: It excels at finding the common denominator. In a multicultural team where some members practice direct feedback and others avoid confrontation entirely, AI can serve as a neutral translator—reformulating, synthesizing, and smoothing out cultural friction. 3. How We Persuade: It’s No Longer About the Argument. It’s About the Person Making It AI produces inductive responses: examples, bullet points, concrete cases. This results-first logic is gradually permeating cultures that traditionally valued deductive reasoning, like in France, for example, where the art of the dissertation (thesis, antithesis, synthesis) was a deep cultural marker. Presentations are getting shorter and more pragmatic. But the real shift goes beyond a simple victory of American-style storytelling over European-style argumentation. What is actually happening is that human embodiment is becoming the primary source of persuasion. When anyone can produce a well-structured argument in 10 seconds, formal argumentative quality stops being differentiating. What convinces people is presence, authenticity, and the personal commitment of the person speaking. 4. How We Lead: From the Lone Expert to the Collective Orchestrator The flattening of knowledge access generated by AI undermines leadership models built on the hoarding of expertise. The manager whose authority derived from mastery of a technical domain sees that competitive advantage eroding. Also, as AI becomes more pervasive, the very source of leadership becomes structurally more collective. AI models are trained on aggregated human work. They are, in a sense, the distillation of millions of anonymous contributions. To use AI is to mobilize a collective intelligence that no single person authored. This should dismantle the myth of the lone brilliant leader. Hence, the leadership of tomorrow may be more about collective discernment and knowing what to do with the output. 5. How We Decide: When the Algorithm Recommends, Do We Still Really Choose? AI compresses decision-making time. In seconds, it produces an analysis, a comparison, a recommendation. And increasingly, we rely on algorithmic recommendations, like HR scoring systems, sales prioritization tools, and project management assistants. Many decisions are made on our behalf. Often, we endorse them without examining them. In cultures that value collective consensus-building before any decision is made, this delegation can feel like a welcome relief. In cultures where strong unilateral decision-making is a mark of leadership, it produces a strange dispossession: The decisive executive finds himself rubber-stamping a recommendation he did not construct. Are we actually still deciding? 6. How We Trust: When All Outputs Look the Same, Relationships Become Everything Here is perhaps the most paradoxical reversal. One might have expected AI to strengthen trust based on the quality of work, since now anyone can produce polished, well-structured deliverables. Instead, the opposite is happening. When all outputs look alike, they lose their power to distinguish. Cognitive trust erodes precisely because AI has made it commonplace. What becomes valuable is the affective, the personal relationship, the two-hour lunch, the intimate conversation. Receiving a proposal that is manifestly generated without human effort sends a signal: You were not worth my real attention. As AI takes over routine interactions, what remains—genuine attention, real presence—acquires extraordinary value. We all crave sincere human contact. The affective dimension of trust is likely to become more precious than ever. 7. How We Disagree: The Risk of a World Where Everyone Agrees AI models avoid confrontation by design. They don’t flatly contradict. They “offer a complementary perspective.” They “acknowledge the nuance.” This algorithmically engineered softness, repeated at a massive scale, may be reshaping the norms of disagreement. In cultures already inclined to avoid open conflict, AI reinforces the tendency to sidestep. In cultures where direct disagreement is seen as healthy and productive, AI introduces a veneer of diplomatic language that can mask real tensions. The danger is organizations where everyone appears to agree—the humans out of politeness, the AIs out of design—and where real problems never surface. A world of frictionless AI-mediated communication will do away with the friction that makes organizations resilient. 8. How We Relate to Time: When a Two-Hour Response Feels Slow AI responds in seconds. That standard of immediacy, internalized across thousands of interactions, is reshaping our tolerance for human response time. A colleague who takes two hours to reply to an email now seems sluggish. A meeting that “takes time” to build consensus feels inefficient. The AI’s instantaneousness has become the invisible benchmark against which all human pace is judged. Cultures that already organize work sequentially and value strict scheduling are accelerating further. Cultures with a more fluid, relational relationship to time—where adaptability matters more than the clock—face growing pressure to conform to responsiveness standards that are foreign to them. The AI has, in effect, exported one particular cultural relationship to time and made it feel universal. The Overlooked Dimension: Gender and the Digital Matilda Effect The numbers are striking. Women are between 20% and 25% less likely than men to use generative AI tools, according to a Harvard Business School meta-analysis. Women are hesitating because they are calculating the risk of being seen using AI. A study found that when engineers submitted identical AI-assisted code for review, women received competence ratings 13% lower; men, only 6% lower. It’s a sort of digital Matilda effect. The historical Matilda effect is the phenomenon by which women’s intellectual contributions are attributed to their male colleagues. When a woman uses AI, observers tend to assume the tool did the thinking. When a man uses the same tool, he is credited with the strategic intelligence to deploy it well. Women who have spent careers navigating this double standard know how to read the room correctly. In thinking with machines, we are changing our codes, our expectations, our relationships, and our hierarchies. Perhaps it’s still too early to fully comprehend the cultural revolution induced by generative AI. But somewhere between the typos we now leave on purpose and the feedback we no longer dare to give, a deeper transformation is already underway—and we have barely begun to notice it. View the full article