-
Posts
7,177 -
Joined
-
Last visited
Content Type
Profiles
Forums
Blogs
Events
Everything posted by ResidentialBusiness
-
It’s no longer groundbreaking to say that the SEO landscape is evolving. But this time, the shift is fundamental. We’re entering an era where search is no longer just about keywords but understanding. At the core of this shift is vector-based SEO. Optimizing for vectors gives websites a major advantage in search engines and overall web presence. As AI and large language models (LLMs) continue to shape digital experiences, websites that adapt early will stay ahead of the competition. What are vectors? Vectors are a mathematical way for AI to understand and organize information beyond just text. Instead of relying on exact keyword matches, search engines now use vector embeddings – a technique that maps words, phrases, and even images into multi-dimensional space based on their meaning and relationships. Think of it this way: If a picture is worth a thousand words, vectors are how AI translates those words into patterns it can analyze. For SEOs, a helpful analogy is that vectors are to AI what structured data is to search engines – a way to provide deeper context and meaning. How vectors change search By leveraging semantic relationships, embeddings, and neural networks, vector-based search allows AI to interpret intent rather than just keywords. This means search engines can surface relevant results even when a query doesn’t contain the exact words from a webpage. For example, a search for “Which laptop is best for gaming?” may return results optimized for “high-performance laptops” because AI understands the conceptual link. More importantly, vectors help AI interpret content that isn’t purely text-based, which includes: Colloquial phrases (e.g., “bite the bullet” vs. “make a tough decision”) Images and visual content. Short-form videos and webinars. Voice search queries and conversational language. Source: Airbyte This shift has been years in the making. Google has been moving toward vector-based search for over a decade, starting with the Hummingbird update in 2013, which prioritized understanding content over simple keyword matching. You might recall RankBrain, Google’s first AI-powered algorithm from 2015, which paved the way for BERT, MUM, and Microsoft’s enhanced Bing Search – all of which rely on vectorized data to interpret user intent with greater accuracy. At its core, vector-based search represents a fundamental change: SEO is no longer about optimizing for exact words but for meaning, relationships, and relevance. As AI continues to evolve, websites that adapt to this approach will have a significant advantage. Dig deeper: AI optimization: How to optimize your content for AI search and agents How vectors impact your SEO strategy So, what does this mean for SEO? If “content is king” has been the mantra for the past decade, then “content is emperor” might be the new reality. A king rules over one kingdom, but an emperor governs many. Similarly, making your content readable to AI doesn’t just improve search engine visibility. It makes your website discoverable across a broader range of AI-driven tools that generate answers to user queries. Practically speaking, there are a few key ways SEOs should adjust their approach to keep websites future-ready. Here are three strategies to start with. From content strategy and keyword research to semantic topic modeling Search volume and keyword difficulty will remain key metrics for now. However, AI tools can provide deeper insights – such as identifying the entities and topics Google associates with your competitors’ content. Instead of just checking keyword volume, analyze the top-ranking pages using NLP tools to see how they structure their topics. Adjust your content briefs to cover semantically related topics, not just one keyword/variations of that keyword. From content optimization to intent matching and semantic SEO Traditional SEO prioritizes exact match keywords and their variations, while AI-driven optimization focuses on aligning with search intent. This means you’ll want to: Run your content through Google’s NLP API to see which topics/entities it detects and compare with competitors that may be ranking better than you. Optimize existing content not only to add keywords, but to add missing context and answer related user queries, by using AlsoAsked and AnswerThePublic. Source: “How to use 12 micro intents for SEO and content journey mapping,” Olaf Kopp From SERP and ranking predictions to AI-based performance forecasting Traditionally, site changes required weeks or months to assess ranking impact. Now, AI can predict performance using vector analysis, giving you another data point for smarter decision-making. Before publishing, paid AI tools like Clearscope or MarketMuse can score your content against high-performing pages. (For smaller projects, free tools like Google Cloud NLP demo offer similar insights.) Use a paid tool like SurferSEO’s SERP Analysis or Outranking.io’s free plan to prioritize content updates based on their likelihood to rank. How vectors don’t change SEO strategy We’re not reinventing the wheel. AI still relies on many of the same principles as traditional SEO. Even if you’re not ready to fully integrate vector-based strategies, you can still optimize your site with them in mind. Great content matters above all else Comprehensive, intent-focused content remains essential for both users and AI, and its importance will only grow. If you haven’t already structured your pages around user intent, now is the time. Write in natural language, focusing on fully answering user queries. Ensure your pages pass the blank sheet of paper test (i.e., they provide unique value on their own). Include synonyms, related phrases, and different ways users might phrase questions. Technical SEO gives AI the roadmap it needs Search engines – and the AI models behind them – still rely on clear signals to understand and rank content effectively. It stands to reason that the use of many of these signals will remain consistent, at least for now. Use structured data to give search engines and AIs more context about the content they’re analyzing. Craft an internal link strategy that makes sense to the average person and demonstrates strong semantic connections between your pages. Dig deeper: Optimizing for AI search: Why classic SEO principles still apply What’s next? As search engines increasingly rely on AI and LLMs, SEO is shifting away from a sole focus on keywords and toward the broader, more intricate concept of meaning. AI systems interpret meaning through vectors, leveraging semantic relationships, embeddings, and neural networks. You can prepare for this shift by optimizing for vector-based search focusing on user intent, content depth, and semantic connections. AI may be the new frontier, but those who embrace change early have the greatest opportunity to drive innovation and shape the future. View the full article
-
Microsoft Bing Search's Copilot Answers now can show maps and local results, including local ads. The AI answer has a map on the left side and then on the right side it has local organic and paid listings with the company name, address, phone, website and directions - plus a photo. View the full article
-
Recently, I overheard a conversation at a local coffee shop: “Thank god for the new administration and finally taking a stand against DEI,” said one of the men to another, as they sipped their coffee. “It’s ridiculous and unfair, completely ruining work. We can finally get back to business.” I leaned in a bit further to try and listen in as I paid for my Earl Gray tea. “Well . . . I’m not sure that’s entirely true,” the other man said hesitating. “I think that . . . ” “Finally, we can get back to raising standards,” the other individual interrupted. “It’s about time! By the way, are you going to the game next week?” The other individual looked uncomfortable as the conversation swiftly shifted in a completely different direction. While I was done paying, and also done eavesdropping, I left knowing that what I heard in this local coffee shop was not an isolated conversation. The backlash against diversity, equity, and inclusion is playing out on the national and world stage almost every single day. And the backlash is also taking place on much smaller stages, in conversations in our conference rooms and in our hallways, amongst colleagues loudly and in whispers in our workplaces. And in these conversations, there’s an opportunity to talk and educate each other about what diversity, equity, and inclusion is and what diversity, equity, and inclusion is not. Here are three of the most common statements I am hearing from individuals for the case against diversity, equity, and inclusion, and here’s how we can debunk these statements and continue to help educate each other on what is true and what is not. False argument against DEI: We lower our standards when it comes to talent Diversity, equity, and inclusion is not about lowering our standards; diversity, equity, and inclusion is about setting fair and equitable standards on how we evaluate all talent. The term “DEI hire” is being used to make us believe that we have lowered standards by hiring individuals from different backgrounds and different lived experiences. In reality, “DEI hire” is a harmful and a hurtful phrase that leads many to believe that someone was handed a job simply because they may “look different” or “be different” or are a “quota hire.” And it is increasingly becoming an acceptable way to discredit, demoralize, and disrespect leaders of color. One of the key outcomes of diversity, equity, and inclusion is creating standardized processes on how we hire talent, and also on who we choose to develop and promote. This includes using software tools like Greenhouse, which helps you ensure that every candidate for a role meets with the same set of interviewers, that interview questions are aligned in advance, and that there’s a way to evaluate and score the interviews and debrief together as an interview team. Otherwise, we fall prey to our biases and may hire people who look like us, think like us, and act like us, or simply hire them because we really just like them. When it comes to how we develop and promote talent, software tools like Lattice help us ensure we set clear and reasonable goals for all, and not just some employees. We can then track progress in weekly meetings, we can give and receive coaching and feedback, and we can have a consistent framework when we evaluate talent during performance review time. And how we evaluate talent is also then connected to how we compensate individuals, and ultimately who we chose to promote. Without these standardized processes, we may end up giving better performance reviews and more money to those who are the most vocal, who spend the most time managing up to us, and who we just find ourselves having more in common with. Diversity, equity, and inclusion efforts help us raise standards and make sure we are getting the best out of our talent. False argument against DEI: It distracts ourselves from driving revenue Diversity, equity, and inclusion does not distract us from leading our businesses; in fact, diversity, equity, and inclusion is a driver of the business. It’s not a separate initiative that sits apart from the business; it should be integrated into everything we do in our workplaces. These efforts not only help us ensure that we get the best out of our talent, but it also ensures we are able to best serve our customers. According to Procter & Gamble, the buying power of the multicultural consumer is more than $5 trillion. Procter & Gamble reminds us that it’s no longer multicultural marketing; it’s in fact mainstream marketing. There is growth to be had when we ensure we connect and authentically serve not just the multicultural consumer, but also veterans, individuals with disabilities, the LGBTQ+ community, and many more communities. Understanding their consumer needs and how your businesses’ products and services can surprise and delight them, and enhance the quality of their lives, is an untapped competitive advantage. Companies like E.L.F. understand this, with a strong focus on diversity, equity, and inclusion efforts that have paid off: It has posted 23 consecutive quarters of sales growth. Over the past five years, the company has also seen its stock increased by more 700%. In contrast, since Target announced a roll back on its diversity, equity, and inclusion efforts, it’s experienced a decline in sales. Black church leaders are now calling on their congregations to participate in a 40 day boycott of Target. Black consumers have $2 trillion in buying power, setting digital trends and engagement. “We’ve got to tell corporate America that there’s a consequence for turning their back on diversity,” said Bishop Reginald T. Jackson, to USA Today. “So let us send the message that if corporate America can’t stand with us, we’re not going to stand with corporate America.” False argument against DEI: An inclusive work environment only benefits a few Diversity, equity, and inclusion is not about creating an inclusive environment for a select few. Diversity, equity, and inclusion is about creating workplaces where we all have an opportunity to reach our potential and help our companies reach their potential. In my book, Reimagine Inclusion: Debunking 13 Myths to Transform Your Workplace, I tackle the myth that diversity, equity, and inclusion processes and policies only have a positive effect on a certain group of individuals. I share “The Curb-Cut Effect” which is a prime example of this. In 1972, faced with pressure from activists advocating for individuals with disabilities, the city of Berkeley, California, installed its first official “curb cut” at an intersection on Telegraph Avenue. In the words of a Berkeley advocate, “the slab of concrete heard ‘round the world.” This not only helped people in wheelchairs. It also helped parents pushing strollers, elderly with walkers, travelers wheeling luggage, workers pushing heavy carts, and the curb cut helped skateboarders and runners. People went out of their way and continued to do so, to use a curb cut. “The Curb-Cut Effect” shows us that one action targeted to help a community ended up helping many more people than anticipated. So, in our workplaces, policies like flexible work hours and remote work options, parental leave and caregiver assistance, time off for holidays and observances, adaptive technologies, mental health support, accommodations for individuals with disabilities, and more have a ripple effect and create workplaces where everyone has an opportunity to thrive. Don’t fall for the rhetoric against “DEI” being exclusive, unfair, or a distraction. The goal of diversity, equity, and inclusion efforts has always been about leveling the playing field and ensuring we are creating workplaces where each and everyone of us have an opportunity to succeed. View the full article
-
A year or more ago, Google added a toggle switch to use Gemini directly in the iOS Google Search app. Well, that was removed this week and you now need to go to the native Gemini app to use Gemini, you cannot use the Google app.View the full article
-
The notion of authenticity in the movies has moved a step beyond the merely realistic. More and more, expensive and time-consuming fixes to minor issues of screen realism have become the work of statistical data renderings—the visual or aural products of generative artificial intelligence. Deployed for effects that actors used to have to create themselves, with their own faces, bodies, and voices, filmmakers now deem these fixes necessary because they are more authentic than what actors can do with just their imaginations, wardrobe, makeup, and lighting. The paradox is that in this scenario, “authentic” means inhuman: The further from actual humanity these efforts have moved, the more we see them described by filmmakers as “perfect.” Is perfect the enemy of good? It doesn’t seem to matter to many filmmakers working today. These fixes are designed to be imperceptible to humans, anyway. Director Brady Corbet’s obsession with “perfect” Hungarian accents in his Oscar-nominated architecture epic, The Brutalist, is a case in point. Corbet hired the Ukraine-based software company Respeecher to enhance accents by using AI to smooth out vowel sounds when actors Adrien Brody and Felicity Jones (American and British, respectively) speak Hungarian in the film. Corbet said it was necessary to do that because, as he told the Los Angeles Times, “this was the only way for us to achieve something completely authentic.” Authenticity here meant integrating the voice of the film’s editor, Dávid Jánsco, who accurately articulated the correct vowel sounds. Jánsco’s pronunciation was then combined with the audio track featuring Brody and Jones, merging them into a purportedly flawless rendition of Hungarian that would, in Corbet’s words in an interview with GQ, “honor the nation of Hungary by making all of their off-screen Hungarian dialogue absolutely perfect.” The issue of accents in movies has come to the fore in recent years. Adam Driver and Shailene Woodley were, for instance, criticized for their uncertain Italian accents in 2023’s Ferrari. Corbet evidently wanted to make sure that would not happen if any native Hungarian speakers were watching The Brutalist (few others would notice the difference). At times, Brody and Jones speak in Hungarian in the film, but mostly they speak in Hungarian-accented English. According to Corbet, Respeecher was not used for that dialogue. Let’s say that for Corbet this will to perfection, with the time and expense it entailed, was necessary to his process, and that having the voice-overs in translated Hungarian-accented English might have been insultingly inauthentic to the people of Hungary, making it essential that the movie sound, at all times, 100% correct when Hungarian was spoken. Still, whether the Hungarian we hear in The Brutalist is “absolutely perfect” is not the same as it being “completely authentic,” since it was never uttered as we hear it by any human being. And, as it turns out, it was partially created in reaction to something that doesn’t exist. In his interview with the Los Angeles Times, Corbet said that he “would never have done it any other way,” recounting when he and his daughter “were watching North by Northwest and there’s a sequence at the U.N., and my daughter is half-Norwegian, and two characters are speaking to each other in [air quotes] Norwegian. My daughter said: ‘They’re speaking gibberish.’ And I think that’s how we used to paint people brown, right? And, I think that for me, that’s a lot more offensive than using innovative technology and really brilliant engineers to help us make something perfect.” But there is no scene in Alfred Hitchcock’s 1959 film North by Northwest set at the United Nations or anywhere else in which two characters speak fake Norwegian or any other faked language. Furthermore, when Corbet brings in the racist practice of brownface makeup that marred movies like 1961’s West Side Story, he is doing a further disservice to Hitchcock’s film. The U.N. scene in North by Northwest features Cary Grant speaking with a South Asian receptionist played by Doris Singh, not an Anglo in brownface. Corbet’s use of AI, then, is based on something that AI itself is prone to, and criticized for: a “hallucination” in which previously stored data is incorrectly combined to fabricate details and generate false information that tends toward gibberish. While the beginning of Hitchcock’s Torn Curtain (1966) is set on a ship in a Norwegian fjord and briefly shows two ship’s officers conversing in a faked, partial Norwegian, Corbet’s justification was based on a false memory. His argument against inauthenticity is inauthentic itself. AI was used last year in other films besides The Brutalist. Respeecher also “corrected” the pitch of trans actress Karla Sofía Gascón’s singing voice in Emilia Pérez. It was used for blue eye color in Dune: Part Two. It was used to blend the face of Anya Taylor-Joy with the actress playing a younger version of her, Alyla Browne, in Furiosa: A Mad Max Saga. Robert Zemeckis’s Here, with Tom Hanks and Robin Wright playing a married couple over a many-decade span, deployed a complicated “youth mirror system” that used AI in the extensive de-agings of the two stars. Alien: Romulus brought the late actor Ian Holm back to on-screen life, reviving him from the original 1979 Alien in a move derided not only as ethically dubious but, in its execution, cheesy and inadequate. It is when AI is used in documentaries to re-create the speech of people who have died that is especially susceptible to accusations of both cheesiness and moral irresponsibility. The 2021 documentary Roadrunner: A Film About Anthony Bourdain used an AI version of the late chef and author’s voice for certain lines spoken in the film, which “provoked a striking degree of anger and unease among Bourdain’s fans,” according to The New Yorker. These fans called resurrecting Bourdain that way “ghoulish” and “awful.” Dune: Part Two [Photo: Warner Bros. Pictures] Audience reactions like these, though frequent, do little to dissuade filmmakers from using complicated AI technology where it isn’t needed. In last year’s documentary Endurance, about explorer Ernest Shackleton’s ill-fated expedition to the South Pole from 1914 to 1916, filmmakers used Respeecher to exhume Shackleton from the only known recording of his voice, a noise-ridden four-minute Edison wax cylinder on which the explorer is yelling into a megaphone. Respeecher extracted from this something “authentic” which is said to have duplicated Shackleton’s voice for use in the documentary. This ghostly, not to say creepy, version of Shackleton became a selling point for the film, and answered the question, “What might Ernest Shackleton have sounded like if he were not shouting into a cone and recorded on wax that has deteriorated over a period of 110 years?” Surely an actor could have done as well as Respeecher with that question. Similarly, a new three-part Netflix documentary series, American Murder: Gabby Petito, has elicited discomfort from viewers for using an AI-generated voice-over of Petito as its narration. The 22-year-old was murdered by her fiancé in 2021, and X users have called exploiting a homicide victim this way “unsettling,” “deeply uncomfortable,” and perhaps just as accurately, “wholly unnecessary.” The dead have no say in how their actual voices are used. It is hard to see resurrecting Petito that way as anything but a macabre selling point—carnival exploitation for the streaming era. Beside the reanimation of Petito and the creation of other spectral voices from beyond the grave, there is a core belief that the proponents of AI enact but never state, one particularly apropos in a boomer gerontocracy in which the aged refuse to relinquish power. That belief is that older is actually younger. When an actor has to be de-aged for a role, such as Harrison Ford in 2023’s Indiana Jones and the Dial of Destiny, AI is enlisted to scan all of Ford’s old films to make him young in the present, dialing back time to overwrite reality with an image of the past. Making a present-day version of someone young involves resuscitating a record of a younger version of them, like in The Substance but without a syringe filled with yellow serum. When it comes to voices, therefore, it is not just the dead who need to be revived. Ford’s Star Wars compatriot Mark Hamill had a similar process done, but only to his voice. For an episode of The Mandalorian, Hamill’s voice had to be resynthesized by Respeecher to sound like it did in 1977. Respeecher did the same with British singer Robbie Williams for his recent biopic, Better Man, using versions of Williams’s songs from his heyday and combining his voice with that of another singer to make him sound like he did in the 1990s. Here [Photo: Sony Pictures] While Zemeckis was shooting Here, the “youth mirror system” he and his AI team devised consisted of two monitors that showed scenes as they were shot, one the real footage of the actors un-aged, as they appear in real life, and the other using AI to show the actors to themselves at the age they were supposed to be playing. Zemeckis told The New York Times that this was “crucial.” Tom Hanks, the director explained, could see this and say to himself, “I’ve got to make sure I’m moving like I was when I was 17 years old.” “No one had to imagine it,” Zemeckis said. “They got the chance to see it in real time.” No one had to imagine it is not a phrase heretofore associated with actors or the direction of actors. Nicolas Cage is a good counter example to this kind of work, which as we see goes far beyond perfecting Hungarian accents. Throughout 2024, Cage spoke against AI every chance he got. At an acceptance speech at the recent Saturn Awards, he mentioned that he is “a big believer in not letting robots dream for us. Robots cannot reflect the human condition for us. That is a dead end. If an actor lets one AI robot manipulate his or her performance even a little bit, an inch will eventually become a mile and all integrity, purity, and truth of art will be replaced by financial interests only.” In a speech to young actors last year, Cage said, “The studios want this so that they can change your face after you’ve already shot it. They can change your face, they can change your voice, they can change your line deliveries, they can change your body language, they can change your performance.” And he said in a New Yorker interview last year, speaking about the way the studios are using AI, “What are you going to do with my body and my face when I’m dead? I don’t want you to do anything with it!” All this from a man who swapped faces with John Travolta in 1997’s Face/Off with no AI required—and “face replacement” is now one of the main things AI is used for. In an interview with Yahoo Entertainment, Cage shared an anecdote about his recent cameo appearance as a version of Superman in the much-reviled 2023 superhero movie The Flash. “What I was supposed to do was literally just be standing in an alternate dimension, if you will, and witnessing the destruction of the universe. . . . And you can imagine with that short amount of time that I had, what that would mean in terms of what I could convey—I had no dialogue—what I could convey with my eyes, the emotion. . . . When I went to the picture, it was me fighting a giant spider. . . . They de-aged me and I’m fighting a spider.” Now that’s authenticity. View the full article
-
The sky is about to get a lot clearer. NASA’s latest infrared space telescope, SPHEREx—short for Spectro-Photometer for the History of the Universe, Epoch of Reionization, and Ices Explorer—will assemble the world’s most complete sky survey to better explain how the universe evolved. The $488 million mission will observe far-off galaxies and gather data on more than 550 million galaxies and stars, measure the collective glow of the universe, and search for water and organic molecules in the interstellar gas and dust clouds where stars and new planets form. The 1107-lb., 8.5 x 10.5-foot spacecraft is slated to launch March 2 at 10:09 pm (ET) aboard a SpaceX Falcon 9 rocket from Vandenberg Space Force Base in California. (Catch the launch on NASA+ and other platforms.) From low-Earth orbit, it will produce 102 maps in 102 infrared wavelengths every six months over two years, creating a 3D map of the entire night sky that glimpses back in time at various points in the universe’s history to fractions of a second after the Big Bang nearly 14 billion years ago. Onboard spectroscopy instruments will help determine the distances between objects and their chemical compositions, including water and other key ingredients for life. SPHEREx Prepared for Thermal Vacuum Testing [Photo: NASA/JPL-Caltech/BAE Systems] Mapping how matter dispersed over time will help scientists better understand the physics of inflation—the instantaneous expansion of the universe after the Big Bang and the reigning theory that best accounts for the universe’s uniform, weblike structure and flat geometry. Scientists hypothesize the universe exploded in a split-second, from smaller than an atom to many trillions of times in size, producing ripples in the temperature and density of the expanding matter to form the first galaxies. “SPHEREx is trying to get at the origins of the universe—what happened in those very few first instances after the Big Bang,” says SPHEREx instrument scientist Phil Korngut. “If we can produce a map of what the universe looks like today and understand that structure, we can tie it back to those original moments just after the Big Bang.” [Photo: BAE Systems/Benjamin Fry] SPHEREx’s approach to observing the history and evolution of galaxies differs from space observatories that pinpoint objects. To account for galaxies existing beyond the detection threshold, it will study a signal called the extragalactic background light. Instead of identifying individual objects, SPHEREx will measure the total integrated light emission that comes from going back through cosmic time by overlaying maps of all of its scans. If the findings highlight areas of interest, scientists can turn to the Hubble and James Webb space telescopes to zoom in for more precise observations. To prevent spacecraft heat from obscuring the faint light from cosmic sources, its telescope and instruments must operate in extreme cold, nearing—380 degrees Fahrenheit. To achieve this, SPHEREx relies on a passive cooling system, meaning no electricity or coolants, that uses three cone-shaped photon shields and a mirrored structure beneath them to block the heat of Earth and the Sun and direct it into space. Searching for life In scouting for water and ice, the observatory will focus on collections of gas and dust called molecular clouds. Every molecule absorbs light at different wavelengths, like a spectral fingerprint. Measuring how much the light changes across the wavelengths indicates the amount of each molecule present. “It’s likely the water in Earth’s oceans originated in a molecular cloud,” says SPHEREx science data center lead Rachel Akeson. “While other space telescopes have found reservoirs of water in hundreds of locations, SPHEREx will give us more than nine million targets. Knowing the water content around the galaxy is a clue to how many locations could potentially host life.” More philosophically, finding those ingredients for life “connects the questions of how `did the universe evolve?’ and `how did we get here?’ to `where can life exist?’ and `are we alone in that universe?’” says Shawn Domagal-Goldman, acting director of NASA’s Astrophysics Division. Solar wind study The SpaceX rocket will also carry another two-year mission, the Polarimeter to Unify the Corona and Heliosphere (PUNCH), to study the solar wind and how it affects Earth. Its four small satellites will focus on the sun’s outer atmosphere, the corona, and how it moves through the solar system and bombards Earth’s magnetic field, creating beautiful auroras but endangering satellites and spacecraft. The mission’s four suitcase-size satellites will use polarizing filters that piece together a 3D view of the corona capture data that helps determine the solar wind speed and direction. “That helps us better understand and predict the space weather that affects us on Earth,” says PUNCH mission scientist Nicholeen Viall. “This`thing’ that we’ve thought of as being big, empty space between the sun and the Earth, now we’re gonna understand exactly what’s within it.” PUNCH will combine its data with observations from other NASA solar missions, including Coronal Diagnostic Experiment (CODEX), which views the inner corona from the International Space Station; Electrojet Zeeman Imaging Explorer (EZIE), which launches in March to investigate the relationship between magnetic field fluctuations and auroras; and Interstellar Mapping and Acceleration Probe (IMAP), which launches later this year to study solar wind particle acceleration through the solar system and its interaction with the interstellar environment. A long journey SPHEREx spent years in development before its greenlight in 2019. NASA’s Jet Propulsion Laboratory managed the mission, enlisting BAE Systems to build the telescope and spacecraft bus, and finalizing it as the Los Angeles’s January wildfires threatened its campus. Scientists from 13 institutions in the U.S., South Korea, and Taiwan will analyze the resulting data, which CalTech’s Infrared Processing & Analysis Center will process and house, and the NASA/IPAC Infrared Science Archive will make publicly available. [Image: JPL] “I am so unbelievably excited to get my hands on those first images from SPHEREx,” says Korngut. “I’ve been working on this mission since 2012 as a young postdoc and the journey it’s taken from conceptual designs to here on the launcher is just so amazing.” Adds Viall, “All the PowerPoints are now worth it.” View the full article
-
Branded is a weekly column devoted to the intersection of marketing, business, design, and culture. Costco chair Hamilton “Tony” James caused a bit of a stir this week when, in an interview, he mentioned a retail category that’s done surprisingly well for the big-box chain: luxury goods. “Rolex watches, Dom Pérignon, 10-carat diamonds,” James offered as examples of high-end products and brands that have fit into a discount-club model more typically associated with buying staples in bulk. “Affluent people,” he explained, “love a good deal.” Courting that group may be particularly timely right now—and not just for Costco. According to a recent report from research firm Moody’s Analytics, the top 10% of U.S. earners (with household incomes of $250,000 and up) now account for almost 50% of all spending; 30 years ago, they accounted for 36%. Moody’s calculates that spending by this top group contributes about one-third to U.S. gross domestic product. While that reflects a serious squeeze further down the income ladder, where the price of eggs and other basics remain high, it also suggests that the comparatively well-off not only have money to spend, but they’re spending it. As of September, the affluent had increased their spending 12% over the prior year, while working-class and middle-class households had spent less. That’s having a clear effect on such sectors as travel, where the affluent have always been part of the mix but are now even more important. Delta recently reported that premium ticket sales are up 8%, much more than main-cabin sales. Bank of America found the most affluent 5% of its customers spent over 10% more on luxury goods than a year ago. “They’re going to Paris and loading up their suitcases with luxury bags and shoes and clothes,” BofA Institute senior economist David Tinsley told the Wall Street Journal. For discounters and dollar stores targeting lower-income consumers, this has meant more competition for fewer dollars spent. (Dollar stores have struggled, and the discount chain Big Lots has filed for bankruptcy.) Addressing that challenge by courting the wealthy isn’t an easy move, but Costco isn’t alone in trying. Walmart CFO John David Rainey told Fox Business that the retail giant has expanded its selection of high-end Apple products, Bose headphones, and other items “sought after by more-affluent customers,” as part of a stab at “upleveling” the Walmart brand. Of course, it may not be easy for most bargain-oriented brands to swiftly pivot, but Costco does seem to have more of a track record of pushing a more upscale-friendly element to its image. James noted that Costco has long counted affluent shoppers among its members—36% of them have incomes of $125,000 and higher, according to consumer-data firm Numerator. A Coresight analysis from a couple years ago found that Costco customers have higher average incomes than those who shop at rival Sam’s Club, and that’s reflected in its brand and product variety. “Costco’s reputation for serving that somewhat higher-income demographic makes the chain more attractive to brands that target those shoppers as well,” Morningstar analyst Zain Akbari told CNBC. And of course it doesn’t hurt that the chain has famously been selling gold and platinum bars. (It’s also worth noting that Costco’s business is doing well in general, not just with high-end customers—and its reputation seems to have benefited from its reaffirmed commitment to DEI as a sound business practice.) “We’ve always known we could move anything in volume if the quality was good and the price was great,” Costco’s James said, and that includes higher-end items that might seem like a stretch for a price-focused warehouse chain. In fact, he argued, it’s a natural part of the Costco brand: Both the company and its fans like to talk about the “treasure hunt”-feel of finding something unexpected that’s not necessarily cheap, but a bargain. “We’re not interested in selling just anything at a low price,” James added. “If someone wants to buy a $500 TV for $250 at Costco, we want to sell them a $1,000 TV for $500 instead. We’re always trying to find better items to sell to members, giving them a great deal. We’re by no means a dollar store.” It (almost) goes without saying that from a macro perspective, the massive wealth and spending-power imbalance underlying Moody’s analysis points to potential problems that shifts in retail strategy won’t solve: Consumer debt in delinquency is rising, and the whole economy is vulnerable if splurging by the affluent were to plummet. So while retailers will likely continue to chase after wealthier customers, the majority of consumers are left to “treasure hunt” for reasonably priced eggs. View the full article
-
In 2021, Eugene Kashuk was looking for a new venture. The Ukrainian entrepreneur realized in the wake of the pandemic that there was a large gap in education. Students were lagging behind, particularly in math. Kashuk started Brighterly, a platform that connects math teachers from all across the globe with students in the United States for private tutoring. Brighterly offers private lessons for $20 per 45-minute lesson—much cheaper than the average rate of about $40 per hour in the United States. In part, Brighterly is able to keep costs down because it uses AI to generate lessons so teachers are able to use their time to focus on their student instead of coming up with problem sets and exams. Fast Company chatted with Kashuk about Brighterly’s growth story and the role of AI in education. How does Brighterly work? Brighterly founder and CEO Eugene Kashuk [Photo: Brighterly] We are a marketplace that connects teachers with students, but we’re not an ordinary marketplace where there’s no control over quality levels or standardization. We have a pretty sophisticated way of hiring and managing teachers and only accept 3% of applicants. We look for teaching skills and soft skills: 90% of our teachers have prior teaching experience. We also have our own custom methodology and curriculum designed by a team of in-house experts, and then we use AI to generate content. On the other side, we’ve got a platform that connects teachers, parents, and kids and allows for lessons to happen. Lessons themselves are fun, interactive, and gamified, and it all revolves around the school curriculum. It’s a high-impact education where we are able to gain academic results as fast as possible. You mentioned using AI to create lessons. How do you make sure it’s not hallucinating or creating misinformation? We are human driven in terms of the content that we create. There are some processes that you can automate using AI. For example, you need to explain fractions to a kid. The best way to do it is to find a circular object such as a pizza you split into pieces. You can use AI to generate images. You can use AI to generate ideas on what kind of circular object will work the best here and will be engaging and fun for the kids. We don’t ask AI to tell us what to teach. As we see it, you can only use AI to create content that is aligned with what you teach. What do you see as the role of AI in education? When I think of AI technologies replacing human educators, I’m more afraid than excited. I’ll caveat and say the technology evolves so fast that you never know what it will look like in a few years. At this point, I don’t see any potential that AI tutors might replace real human educators. In order for education to work, teachers need to form an emotional bond with their students. Not everything can be covered by logic and algorithms; you need to have that human input to understand what the child really needs. We’re not at a moment where AI can really replace human interaction in education. However, we have great capabilities to generate content or to generate language to help create lessons and personalized assessments, which can be very useful for educators. So you’re saying AI can lighten the load for teachers so they can focus on nourishing the emotional bond with students. And so teachers can also focus on decision-making. You need a human teacher to understand how well a student is doing and assess what they need next. For example, maybe a student seems to grasp a topic but the teacher can sense that they’d benefit from some more repetition. I don’t see at this point how AI can pick up on moments like this. Is there anything else you’d like to add? The education gap isn’t just in math—there’s also a reading problem. We get daily requests for more reading lessons, so we’ll be launching a reading course as well. There was a COVID relief program that allocated resources to cover those gaps, but it’s unclear if it’ll be funded in the future. In the meantime, the knowledge gap is growing. Currently, we only cover elementary and middle school, but we’ll also be launching a high school product to help with that gap. Children are the future and right now existing solutions for the knowledge gap aren’t working. We need more. View the full article
-
At Buffer, we are always looking for ways to help our team work more effectively and creatively. Our latest investment in productivity is a company-wide AI Tools Stipend of $250 per teammate per year. This initiative reinforces our ongoing commitment to supporting innovation while giving our teammates the autonomy to choose tools that best fit their individual workflows. Here's why we're introducing this stipend and how we expect it to impact our work. The evolution of AI in our workflowsOver the past year, we've seen AI tools rapidly transform how many of us work. What started as occasional experimentation with tools like ChatGPT has evolved into more regular integration of AI assistants in various aspects of our workflow — from drafting content and generating ideas to debugging code and analyzing data. Late last year, our Engineering Managers began exploring the idea of offering a stipend specifically for Engineers to experiment with productivity-boosting AI software. We noticed some inconsistencies in reimbursement requests for these tools and aligned on providing structured support for this exploration. The positive impact on our engineering team made it clear that AI tools could benefit everyone at Buffer, regardless of their role. That's why we've decided to extend this stipend to all teammates. Why we're introducing an AI tools stipendThe pace of AI tool development has accelerated dramatically, and we want to empower our team to experiment and find what works best for their specific needs. There are a few key reasons behind this decision: Autonomy and flexibility: Different roles benefit from different AI tools, and we want teammates to have the freedom to choose what works for them rather than prescribing a one-size-fits-all solution.Reducing friction: By providing a dedicated stipend, we remove financial barriers to trying new tools that might significantly improve productivity.Learning together: As we all explore different AI tools, we can share insights and recommendations, collectively building our understanding of how these technologies can best support our work.Staying ahead of the curve: AI technology is evolving rapidly, and we want to ensure our team has access to cutting-edge tools that help us continue delivering exceptional service to our customers.How our AI stipend worksWe've allocated $250 per teammate per year for AI tools. With our current team size of 71 people, this represents a potential additional operating expense of up to $17,750 annually. The stipend offers flexibility, allowing teammates to maintain a yearly subscription to one primary tool, access others occasionally for specific projects, or mix and match several tools throughout the year. We've designed the AI Tools Stipend to be straightforward and flexible: Eligibility: All Buffer teammates are eligible to use the full $250 stipend annually.Covered tools: The stipend covers AI-powered tools such as ChatGPT, Claude, Midjourney, Cursor, Cody, Spiral, Raycast, and similar tools that enhance productivity and creativity.Subscription recommendation: We suggest opting for monthly subscriptions rather than annual commitments to maintain flexibility in trying different tools.Free alternatives: Many AI tools offer free versions, so we recommend exploring those first before committing to a paid plan. Additionally, Gemini Advanced is already available through our Google Workspace.Knowledge sharing: In the spirit of transparency, teammates who use the stipend are encouraged to share the tool they're exploring in our #culture-ai Slack channel so we can all learn together.Learning together: building an AI knowledge baseOne of the most exciting aspects of this program is the opportunity to collectively discover and share best practices for using AI tools. By encouraging discussion in our #culture-ai Slack channel, we hope to: Create a repository of real-world use cases specific to Buffer's workHelp teammates discover tools that might benefit their particular workflowIdentify patterns in how different teams are leveraging AIWe already have some early success stories: Content Team: Experimenting with AI for research and ideation while maintaining our unique voice. The stipend is already having an impact here as our content writer, Tami, wrote an article detailing an experiment with five AI chatbots and how they might fit into creators’ workflows.Engineering Team: Boosting productivity through AI-assisted coding.Customer Advocacy Team: Exploring ways to enhance customer interactions with AI support.Through these shared experiences, we'll continue to refine our understanding of how AI can best support our work at Buffer. Over to youThis stipend is just one part of our broader approach to thoughtfully integrating AI into our work at Buffer. We see AI tools as assistants that can handle routine tasks and provide creative inspiration, freeing up our human team to focus on the work that requires genuine human connection, creativity, and strategic thinking. We’d love to hear your thoughts: Do you have questions about how we're approaching AI adoption at Buffer?Are you using AI tools in your work? Which ones have you found most valuable?As always, we're learning as we go — we believe in the potential of AI to enhance our work while still keeping human creativity, connection, and judgment at the center of what we do. Read more about AI at Buffer🧠 The AI Mindset Shift: How I Use AI Every Day 📚 How 1.2 Million Posts Created with Buffer’s AI Assistant Performed vs Human-Only — and What That Means for Your Content 🤖 How Buffer’s Content Team Uses AI 🛠️ The Story Behind Buffer’s AI: An Interview with Diego Sanchez View the full article
-
“But what is death?” I am sitting down with Katrina Crawford and we are here to talk about the White Lotus Season 3 opening credits. Together with Mark Bashore, Crawford runs the creative studio Plains of Yonder, which has crafted the White Lotus main titles for every season so far. But that question about death wasn’t posed by me. It was posed by her. And it challenges us to reflect on the meaning of death, and the many ways to die. Since White Lotus season 3 premiered on February 16, the internet has been abuzz with theories and criticisms around who died and what the opening sequence means. In response, HBO has said: “You’ll get it soon enough.” So while we wait, we decided to call up Crawford and Bashore so we can dissect one of the most iconic main titles in modern history. The biggest takeaway? Some things can have more than one meaning. [Photo: Courtesy Plains of Yonder] Easter eggs or red herrings? Plains of Yonder has made over a dozen main titles for shows like The Decameron and The Lord of the Rings TV show, Rings of Power. But Crawford says that this title, for White Lotus Season 3, is by far the longest they’ve ever spent developing. While some showrunners don’t consider the main title until the end, Crawford says that Mike White gave the team a whole 10 months to craft the sequence. When they started, Cristobal Tapia de Veer, who wrote the music for previous seasons’ openers, was still composing the new soundtrack. The crew wasn’t even shooting yet. All they had was the script. White Lotus Season 3 has eight episodes. They got the script for the first seven episodes. “We know a lot,” Crawford tells me. But they don’t know who dies. [Photo: Courtesy Plains of Yonder] Crawford combed through those scrips and crafted meticulous profiles for every character, where she tried to understand who these characters are—or who she thinks they are. Once the profiles were complete, she assigned dozens of images to them. Sometimes, she paired a character with an animal (a stoned monkey for the North Carolina mom played by Parker Posy). Sometimes, she crafted a scenario around them (Jason Isaacs’s Timothy Ratliff character appears coiled up in a tree with knives for branches.) These images might intimate a character’s fate or personality, but of course, some interpretations are more literal than others. “Maybe someone is presenting one way, but something else is truth,” she says. [Photo: Courtesy Plains of Yonder] Animal instincts As with every season, animals carried much of the symbolism. “We’ve always found that using animals are a better metaphor than people,” says Bashore. Season 1 was obsessed with monkeys; Season 2 introduced humping goats; Season 3 goes all in on mythological creatures that are half-human, half-beast. (Crawford spent a month poring over Thai mythology books.) The intro opens with a circus of animals, and the first humans to be portrayed are human faces attached to bird bodies. A bit later into the sequence, Lek Patravadi’s Sritala Hollinger character (the hotel owner) appears by a pond, holding a creature that is half human, half bird—perhaps a clue about her mysteriously absent husband. Most of these metaphors are predictably obscure. Amy Lou Wood (who plays Chelsea) appears in the middle of an incriminating scene depicting a leopard that has bitten off a deer’s head with two foxes bearing witness. Which animal is she? Meanwhile, Natasha Rothwell, who plays Belinda from Season 1, is portrayed next to a stork staring down at its reflection in the water, while a crocodile is lying in lurk. Will it snap her up in its jaws? [Photo: Courtesy Plains of Yonder] Fiction or reality? That the team spent ten months developing the title isn’t so surprising considering the complexity of these characters. But there is one more character in the story, and that is Thailand. Like with the first two seasons, the location plays a key role in the story. To build that sense of place, the team spent ten days filming at three Royal temples in Thailand. “We shot the daylights out of it,” says Bashore, noting that they took over 1,000 photos of patterns, colors, outfits, and of course, those iconic Thai rooflines. “We looked for quick visual cues, patterns that feel iconically Thailand.” [Photo: Courtesy Plains of Yonder] Getting permits to film was no small feat considering the sacred nature of the temples. The team had to coordinate access with HBO, and other logistical challenges meant that the title took longer to make. But could the delay signal something else? Some shows, like Game of Thrones, have created main titles that change from week to week. Was the White Lotus Season 3 main title so challenging to make because the team had to tweak it as the season progresses? Crawford gives a cheeky shrug that neither confirms nor denies it: “We can’t talk about anything you haven’t seen.” What we can talk about is what Crawford calls “the Temple of White Lotus.” Indeed, the temples help anchor the show in Thailand, but the real star of the show is the White Lotus resort, which may have been filmed at the Four Seasons Resort Koh Samui, but remains a fictional place. To emphasize the otherworldly setting, the team broke down the photographs they took into little squares and shuffled them to create new images out of them—like a patchwork that looks and feels real, but ultimately isn’t. “We’re not trying to say this is a real place,” says Bashore. “It’s just a vibe we’re soaking in.” [Photo: Courtesy Plains of Yonder] This vibe ultimately blossoms in the seven worlds the team created for the title sequence. The story begins in the jungle, then unfolds in a village, a temple, a pond (in which we see Sam Nivola’s body afloat), a gloomier forest with ominous snakes coiled around trees, an epic battle scene, and the grand finale—“disaster at sea”—where a throng of men gets swallowed by giant fish. Could this mass killing scene signify more than one death in the show? Crawford’s response? You know it already: “But what is death?” View the full article
-
For more than two decades, users have turned to search engines like Google, typed in a query, and received a familiar list of 10 blue links—the gateway to the wider web. Ranking high on that list, through search engine optimization (SEO), has become a $200 billion business. But in the past two years, search has changed. Companies are now synthesizing and summarizing results into AI-generated answers that eliminate the need to click through to websites. While this may be convenient for users (setting aside concerns over hallucinations and accuracy) it’s bad for businesses that rely on search traffic. One such business, educational tech firm Chegg, has sued Google in federal district court, alleging that AI-generated summaries of its content have siphoned traffic from its site and harmed its revenue. Chegg reported a 24% year-on-year revenue decline in Q4 2024, which it partly attributes to Google’s AI-driven search changes. In the lawsuit, the company alleges that Google is “reaping the financial benefits of Chegg’s content without having to spend a dime.” A Google spokesperson responded that the company will defend itself in court, emphasizing that Google sends “billions of clicks” to websites daily and arguing that AI overviews have diversified—not reduced—traffic distribution. “It’s going to be interesting to see what comes out of it, because we’ve seen content creators anecdotally complaining on Reddit or elsewhere for months now that they are afraid of losing traffic,” says Aleksandra Urman, a researcher at the University of Zurich specializing in search engines. Within the SEO industry, anxiety over artificial intelligence overviews has been mounting. “Chegg’s legal arguments closely align with the ethical concerns the SEO and publishing communities have been raising for years,” says Lily Ray, a New York-based SEO expert. “While Google has long displayed answers and information directly in search results, AI overviews take this a step further by extracting content from external sites and rewording it in a way that positions Google more as a publisher than a search engine.” Ray points to Google’s lack of transparency, particularly around whether users actually click on citations in AI-generated responses. “The lack of visibility into whether users actually click on citations within AI overviews leaves publishers guessing about the true impact on their organic traffic,” she says. Urman adds that past research on Google’s featured snippets—which surfaced excerpts from websites—showed a drop in traffic for affected sites. “The claim seems plausible,” she says, “but we don’t really have the evidence to say how the appearance of AI overviews really affects user behavior.” Not all companies are feeling the squeeze, however. Ziff Davis CEO Vivek Shah said on an earnings call that AI overviews had little effect on its web traffic. “AI’s presence remains limited,” he said. “AI overviews are present in just 12% of our top queries.” It remains to be seen whether Chegg is an outlier or a bellwether. Ray, for her part, believes its lawsuit could be a pivotal moment in the fight over AI and SEO. This case will be fascinating to watch,” she says. “Its outcome could have massive implications for millions of sites beyond just Chegg.” View the full article
-
There’s a new website that tracks how much of Project 2025, the 922-page conservative playbook, has already come to fruition under President Donald Trump’s administration. It shows that, in less than two months, more than a third of the right-wing agenda’s objectives have been fulfilled. The site, called Project 2025 Tracker, is broken down into bite-sized sections based on the goals laid out in document. Project 2025 was written by the conservative think tank Heritage Foundation and several members of Trump’s first administration, and although the President distanced himself from the document on the campaign trail, he’s since said he agrees with many parts of it. His administration has already proved itself to be at least partially aligned with the project’s policy goals. While other major media outlets have compiled list-style round-ups of President Trump’s executive orders thus far, the tracker takes a more visual, big-picture approach to understanding how closely Trump’s second term is mapping onto a far-right blueprint—including a Project 2025 progress bar. [Image: Project 2025 Tracker] Breaking down Project 2025 The Project 2025 Tracker is the result of an unplanned collaboration between two Reddit users. It started with Adrienne Cobb, an archaeologist by trade who, outside of work, runs a political and legal news subreddit with 183,000 members called r/Keep_Track. Last month, Cobb decided to make a spreadsheet tracking the Trump administration’s progress on Project 2025 initiatives, which quickly gained traction on her subreddit. “Project 2025 is one of the biggest threats to democracy, to the common good, that we’ve faced,” Cobb says. “It is important that Americans understand what it is and how it may impact their lives. But most people don’t have time to read a 900-page document. So, I set out to read it myself and extract the objectives to make it easier for everyone to digest.” To build the spreadsheet, Cobb sifted through every chapter of Project 2025, each of which “addresses a specific federal agency and recommends how a president can ‘reform’ it in accordance with the Heritage Foundation’s vision,” she says. For every agency section, Cobb identified a series of “objectives,” which she’s defined as any moment when the author of a chapter explicitly calls for an action to be taken. This is to avoid “any subjective reading between the lines,” Cobb says. That includes everything from a suggestion that the Center of Disease Control “should immediately end its collection of data on gender identity” to a directive to reverse the Department of Defense’s “policies that allow transgender individuals to serve in the military” (both of which have since been carried out). The spreadsheet only transformed into a website when fellow Reddit user u/mollynaquafina stumbled across it during their daily spin through the site. “I was scrolling through Reddit, like I do most mornings, and I came across a post from /u/rusticgorilla in the /r/keep_track subreddit,” u/mollynaquafina says. “They posted about the Project 2025 spreadsheet tracker they started. I checked it out and was impressed with the level of detail—with references to specific page numbers and news articles for each objective.” As someone with a background in cybersecurity, u/mollynaquafina says, seeing the information compiled in a spreadsheet “helped to connect the dots in my brain.” Still, they add, others might process information in a visual way. To make the information more transparent, they offered to turn the spreadsheet into a more digestible website. ‘The sheer amount of news each day is staggering’ The first priority of the Project 2025 Tracker, u/mollynaquafina explains, is to highlight just how much progress has already been made in the agenda’s objectives. To that end, the website’s top section is a bright blue “overall progress” bar coupled with a live countdown of time remaining in Trump’s second term. Currently, with 1,423 days left to go, overall Project 2025 progress has already reached 36%. “Trump’s second term has been much more chaotic than his first, which I didn’t think was possible,” Cobb says. “The sheer amount of news each day is staggering. I believe that is the goal, though: to move fast, to overwhelm, to stun. By the time ‘the people’ have a chance to organize against a particular action, the administration has already moved on to the next objective.” As users dig deeper into the site, they can also filter Project 2025 progress by agency, subject (like DEIA or energy), and status. Any objective that has already been carried out is highlighted with a green “completed” bubble and accompanied with links to both a citation in the Project 2025 document and an accompanying news item. U/mollynaquafina has also added a chart and timeline feature to help users take in the holistic picture. “I was very intentional about using factual and neutral language where possible,” U/mollynaquafina says. “My goal is for people who support the Project 2025 objectives to clearly read about them and ask themselves, ‘Is this what I signed up for?’” To keep up with the constant influx of news, Cobb uses what she calls a “low-tech” process of combing through social media, news articles, and federal websites before manually organizing stories by topic in a word-processing document. From there, she compares her notes against the spreadsheet to determine if any news events match up with Project 2025 objectives. The tracker is pulling in around 50,000 to 100,000 visitors each day, with about 10% of traffic originating outside the U.S. Going forward, u/mollynaquafina plans to translate the site into several other languages based on demand. “The community has been very supportive of our effort to track Project 2025 but has also expressed a lot of fear about where the country is headed,” Cobb says. “I think that is totally justified and normal. My hope is that helping people understand Project 2025 will remove some of that powerlessness, remove the fear of the unknown, and spur more collective action to stop its progress.” View the full article
-
Unless you’re at the very top of the food chain in your organization, you report to someone. And that manager is important for your career success. They will evaluate your performance, give you feedback and mentoring, greenlight ideas, and provide support elsewhere in the organization for things you’re doing. Because of all the roles that a supervisor plays for you, it can be stressful when a new person steps in, or you get promoted and start reporting to someone new. There are several ways you can make this transition easier and lay the groundwork for a fruitful relationship with your new boss: Be mindful of the firehose When your supervisor is replaced with someone else, that person is stepping into a new role. Whenever you take on something new, there is a lot that comes at you those first few weeks. You need to get to know your team, your new responsibilities, your new peer groups. All of that happens while work needs to continue. That’s the reason the information dump you face when you first start a role is often referred to as “drinking from the firehose.” If your new boss is drinking from that hose, then you want to provide them with information slowly. Invite them to let you know when you can have a chance to meet and talk, recognizing that it may take them a week or more to get settled in before they’re ready to have an in-depth conversation. As your supervisor gets settled, you can expand the amount of information you provide. Giving them that information when they’re ready to receive it will greatly increase the likelihood it gets read at all. Provide an introduction You do want to introduce yourself in stages. Start with a quick, one-paragraph introduction to share who you are, how long you’ve been with the organization, and your role on the team. If there are one or two critical things that your supervisor needs to call on you for, then highlight those as well. After that, you can add more information about your broader responsibilities. Highlight projects that you’re working on that may need your supervisor’s input or attention. You might consider creating a cheat sheet for your new boss that lists the things you’re working on, your best contact information, and your key strengths and weaknesses. That way, your supervisor knows when to call on you and the kinds of projects where it might be helpful to include you. After the initial rush, you can add details about projects that don’t need immediate attention and other projects you have been involved in so that your supervisor can recognize the contributions you’ve made. Talk about expectations Your previous supervisor not only knew what you had done but also had a sense of your career trajectory, interests, and goals. Hopefully this person was also actively engaged in helping you grow and providing opportunities for you to move forward with your career goals. Your new boss doesn’t know any of these things about you. If you have expectations for how you want to be treated and what you would like your supervisor to do to help you succeed in your role and advance, you need to talk about those expectations. It’s also helpful for you to discuss any weaknesses that you would like help to address. It is common to want to start your interactions with your supervisor by focusing on your strengths. And you should do that. But you might also be tempted to hide your weaknesses. Instead, it can be useful to be up front with those aspects of your work that you are still developing. It can ensure that you get help when you need it and that you’re considered for professional development opportunities. Ask for what you need Consistent with highlighting your weaknesses, you should generally ask your new boss for what you need from them. If you like to get feedback in the middle of projects, ask for it. If you need more information about the strategy behind a project or plan, ask. Nobody is a mind reader. And the less well someone knows you, the less able they are to anticipate what you need from them. Rather than hoping your new boss will give you everything you need, you’ll have to ask for it. Usually, a supervisor will be glad to have some guidance about how you’d like to engage with them. Asking for what you need is the best way to get the relationship with your new supervisor off to a good start. View the full article