If any one wanna try it in browser can try implement it on our chrome extension code base in typescript: https://github.com/nanobrowser/nanobrowser, we support browser use and re wrote its code in typescript, not yet supporting workflow use and love to hear how it works from community!
we built a chrome extension supporting browser use and can run locally in ur browser: https://github.com/nanobrowser/nanobrowser, feel free to implement workflow use on top of our code base in typescript. Love to learn how it works!
we built a chrome extension supporting browser use: https://github.com/nanobrowser/nanobrowser, feel free to implement workflow use on top of our code base in typescript. Love to learn how it works!
We completely rewrote the DOM processing part of browser-use with Typescript, built a multi-agent system and implemented the extension frontend with React, Vite and Typescript.
Quick clarification. Low domain knowledge is okay for those who don't have experience and don't know what to say, like Alexa. High domain is somebody who has expertise with a specialized workflow.
So, they will rely on voice commands for recognition, not natural language. Often one to two words to set a chain of tasks in motion. Think of having to control your entire computer, including navigating by voice. That would be very exhausting and inefficient through natural language. There needs to be a hybrid solution that can leverage low domain natural language, but also high domain command-based recognition. I cannot overstate how important of low latency between the beginning of a command and an action produced. High latency means a big cognitive load and not to mention just inefficiency.
There's a lot of overlap between UI automation and accessibility control tools. However, UIA automation has always been a slow process simply because the stack doesn't have the demand from devs for low latency.
It's a difference between having an independent agent do something on your behalf, not caring how long it takes, versus you waiting for a aynchronous task to be completed.
appreciate clarification. The low domain vs high domain distinction is spot-on:Latency kills expert workflows. keeping this in mind when integrating/designing voice recognition and more accessibility control options.
We are grateful for everyone contributing in open source, and would love for you to join our community to help making open source AI automation better!
Even if you're not interested in contributing code, you can help by:
-Trying out Nanobrowser and sharing your feedback.
-Suggesting new prompts and use cases.
-Helping us build a comprehensive evaluation framework.
-Joining our Discord community: https://discord.gg/NN3ABHggMK
We know it's early days, and there's still a lot to improve. We appreciate your support in such early times! Thanks for checking it out! We're excited to see what you build with Nanobrowser.