Woolworths has introduced a partnership with Google to include agentic synthetic intelligence into its “Olive” chatbot, beginning in Australia later this 12 months.
Till now, Olive has largely answered questions, resolved issues and directed buyers to info.
Quickly, Olive will be capable to do extra: planning meals, decoding handwritten recipes, making use of loyalty reductions and inserting recommended objects instantly right into a buyer’s on-line purchasing basket.
Woolworths
Woolworths says Olive won’t full purchases routinely, and clients will nonetheless have to approve and pay for orders.
This distinction is essential, however dangers understating what’s truly altering. By the point a consumer reaches the checkout, most of the substantive selections about what to purchase might have already got been formed by the system.
From helper to resolution maker
Essentially the most vital change for buyers is how selections can be made through the purchasing course of – and who makes them.
Google describes its new system as a “proactive digital concierge” that understands buyer intent, causes by way of multi-step duties, and executes actions.
Main United States retailers, together with Walmart, Kroger and Lowe’s, are adopting the identical expertise. The transfer varieties a part of a broader technique by Google to advertise agent-based commerce throughout retail.
In sensible phrases, if Woolworths buyers give their permission, the brand new Google Gemini model of Olive will more and more assemble purchasing baskets autonomously.
For instance, a buyer who uploads a photograph of a handwritten recipe may obtain a accomplished listing of components, reflecting product availability and reductions.
Alternatively, a buyer who asks for a meal plan may obtain a ready-made basket based mostly on previous preferences, present promotions and native inventory ranges.
This basically adjustments the function of the consumer.
As a substitute of actively deciding on merchandise by way of searching and comparability, buyers will more and more assessment and approve alternatives made for them. Resolution-making shifts away from the person in direction of the system.
This delegation might seem minor when thought of in isolation. Over time, nevertheless, repeated delegation shapes habits, preferences and spending patterns. That’s the reason this new change deserves cautious scrutiny.
Nudging by design
Woolworths presents Olive’s expanded function as a sensible comfort to avoid wasting effort and time, whereas growing personalisation. These claims aren’t incorrect, however they obscure an essential level.
Agent-based purchasing methods are designed to nudge behaviour in ways in which differ markedly from conventional promoting.
When Olive highlights discounted merchandise or promotional affords for a consumer, it doesn’t depend on impartial standards. As a substitute, its priorities mirror pricing methods, promotional priorities and business relationships – not an goal evaluation of the patron’s pursuits.
As soon as such judgements are embedded inside an AI system that guides purchasing selections, nudging turns into a part of the construction of alternative, relatively than a visual layer positioned on prime of it.
It is a significantly highly effective type of affect. Conventional promoting is recognisable. Consumers know when they’re being persuaded and may low cost or ignore it.
Algorithmic nudging, in contrast, operates upstream. It shapes which choices are surfaced, mixed, or omitted earlier than the consumer encounters them. Over time, this affect turns into routine and troublesome to detect.
Agent-based purchasing additionally means AI does the searching, evaluating costs and weighing alternate options for us. Consumers are more and more offered with curated outcomes that invite acceptance, relatively than deliberation.
As fewer choices are made seen and fewer trade-offs are explicitly offered, comfort begins to interchange knowledgeable alternative.
For these causes, it will be mistaken to deal with agent-led purchasing as worth impartial. Techniques designed to extend loyalty and income shouldn’t routinely be assumed to behave in one of the best pursuits of customers, even once they ship real comfort.
Unresolved knowledge privateness questions
Information privateness is a good larger concern.
Grocery purchasing reveals excess of model desire. Meal planning can disclose well being situations, dietary restrictions, cultural practices, non secular observance, household composition and monetary pressures. When an AI system manages these duties, home life turns into legible to the platform that helps it.
Google has acknowledged buyer knowledge utilized in its system will not be used to coach fashions and that strict security requirements apply.
These assurances are essential, however they don’t resolve all issues. It’s not but clear how lengthy family knowledge is retained, the way it’s aggregated, or how insights from such knowledge are used elsewhere.
Consent affords restricted safety on this context. It’s usually granted as soon as, whereas profiling and optimisation proceed over time. Even with out direct knowledge sharing, inferences drawn from family behaviour can form system efficiency and design.
These privateness dangers don’t rely upon misuse or knowledge breaches. They come up from the rising intimacy of information used to form behaviour, relatively than merely document it.
Comfort shouldn’t finish the dialog
For a lot of households, Olive’s expanded capabilities will save time, scale back friction and enhance the purchasing expertise.
However when AI strikes from help to motion, it reshapes how decisions are made and the way a lot company folks quit.
This shift ought to immediate a broader dialogue about the place comfort ends and client autonomy begins. When AI methods begin making on a regular basis selections, we should ask whether or not customers retain significant management over their decisions.
Transparency about how suggestions are generated, limits on business incentives shaping agent behaviour, and limits on family knowledge use needs to be handled as baseline expectations, not elective safeguards.
With out such scrutiny, agent-led purchasing dangers quietly reconfiguring client behaviour in methods which are troublesome to detect – and even tougher to reverse.![]()
This text is republished from The Dialog underneath a Artistic Commons license. Learn the unique article.
Elevate your perspective with NextTech Information, the place innovation meets perception.
Uncover the most recent breakthroughs, get unique updates, and join with a worldwide community of future-focused thinkers.
Unlock tomorrow’s tendencies as we speak: learn extra, subscribe to our e-newsletter, and turn out to be a part of the NextTech neighborhood at NextTech-news.com

