Google is developing an AI agent that manages daily tasks

Key Takeaways
  • In December, Google might launch “Project Jarvis,” a LAM with Rabbit-inspired features.
  • It is anticipated that the tool will help with a variety of web chores and make its debut alongside the Gemini LLM.
  • To find and fix any flaws, there can be a restricted release for testers.

In order to enable web task automation in Chrome, Google may introduce “Project Jarvis,” its Rabbit-inspired model, in December.

According to The Information, the corporation plans to showcase the computer-using agent concurrently with the debut of its main Gemini large language model (LLM).

Named after J.A.R.V.I.S. from Iron Man, “Project Jarvis” would only function with a web browser, primarily Chrome. According to sources, the program might assist users in automating routine web actions such as taking and evaluating screenshots, clicking buttons, typing text, booking flights, conducting research, and buying online. Whether this is for desktop or mobile devices is not made clear in the text.

According to the study, Jarvis takes “a few seconds” to execute tasks, suggesting that it probably uses the cloud rather than running on-device.

According to reports, Google is thinking about limiting the rollout to testers in order to find and address faults. The information cautions that the firm may alter its December showcase plan for Jarvis.

AI Businesses Use LAMs to Push Boundaries

Booking rooms and making complicated judgments are made possible by a LAM, an AI system that converts human intentions into actions. For real-time responses and strategic planning, LAMs learn from large user activity datasets.

Prominent AI firms are developing LAMs akin to the one detailed in the Google study. For example, Anthropic recently introduced Claude, a chatbot that uses artificial intelligence (AI) to carry out complicated computer tasks on its own. With users’ permission, Claude handles on-screen data and takes action. According to reports, OpenAI is also developing a similar version.

Users will be able to engage with it in relation to the webpages they read, thanks to Microsoft’s Copilot Vision. Next year, it’s anticipated that Apple Intelligence will be able to comprehend screen content and carry out activities across several apps.

Leave a Reply

Your email address will not be published. Required fields are marked *