https://x.com/AndrewYNg/status/1882125891821822398
Our first short course with @AnthropicAI! Building Towards Computer Use with Anthropic. This teaches you to build an LLM-based agent that uses a computer interface by generating mouse clicks and keystrokes. Computer Use is an important, emerging capability for LLMs that will let AI agents do many more tasks than were possible before, since it lets them interact with interfaces designed for humans to use, rather than only tools that provide explicit API access. I hope you will enjoy learning about it!
This course is taught by Anthropic's Head of Curriculum, @Colt_Steele. You'll learn to apply image reasoning and tool use to "use" a computer as follows: a model processes an image of the screen, analyzes it to understand what's going on, and navigates the computer via mouse clicks and keystrokes.
This course goes through the key building blocks, and culminates in a demo of an AI assistant that uses a web browser to search for a research paper, downloads the PDF, and finally summarizes the paper for you.
In detail, you’ll:
- Learn about Anthropic's family of models, when to use which one, and make API requests to Claude
- Use multi-modal prompts that combine text and image content blocks, and also work with streaming responses
- Improve your prompting by using prompt templates, using XML to structure prompts, and providing examples
- Implement prompt caching to reduce cost and latency
- Apply tool-use to build a chatbot that can call different tools to respond to queries
- See all these building blocks come together in Computer Use demo
Please sign up here: https://deeplearning.ai/short-courses/building-towards-computer-use-with-anthropic
Our first short course with @AnthropicAI! Building Towards Computer Use with Anthropic. This teaches you to build an LLM-based agent that uses a computer interface by generating mouse clicks and keystrokes. Computer Use is an important, emerging capability for LLMs that will let AI agents do many more tasks than were possible before, since it lets them interact with interfaces designed for humans to use, rather than only tools that provide explicit API access. I hope you will enjoy learning about it!
This course is taught by Anthropic's Head of Curriculum, @Colt_Steele. You'll learn to apply image reasoning and tool use to "use" a computer as follows: a model processes an image of the screen, analyzes it to understand what's going on, and navigates the computer via mouse clicks and keystrokes.
This course goes through the key building blocks, and culminates in a demo of an AI assistant that uses a web browser to search for a research paper, downloads the PDF, and finally summarizes the paper for you.
In detail, you’ll:
- Learn about Anthropic's family of models, when to use which one, and make API requests to Claude
- Use multi-modal prompts that combine text and image content blocks, and also work with streaming responses
- Improve your prompting by using prompt templates, using XML to structure prompts, and providing examples
- Implement prompt caching to reduce cost and latency
- Apply tool-use to build a chatbot that can call different tools to respond to queries
- See all these building blocks come together in Computer Use demo
Please sign up here: https://deeplearning.ai/short-courses/building-towards-computer-use-with-anthropic
https://x.com/AndrewYNg/status/1882125891821822398
Our first short course with @AnthropicAI! Building Towards Computer Use with Anthropic. This teaches you to build an LLM-based agent that uses a computer interface by generating mouse clicks and keystrokes. Computer Use is an important, emerging capability for LLMs that will let AI agents do many more tasks than were possible before, since it lets them interact with interfaces designed for humans to use, rather than only tools that provide explicit API access. I hope you will enjoy learning about it!
This course is taught by Anthropic's Head of Curriculum, @Colt_Steele. You'll learn to apply image reasoning and tool use to "use" a computer as follows: a model processes an image of the screen, analyzes it to understand what's going on, and navigates the computer via mouse clicks and keystrokes.
This course goes through the key building blocks, and culminates in a demo of an AI assistant that uses a web browser to search for a research paper, downloads the PDF, and finally summarizes the paper for you.
In detail, you’ll:
- Learn about Anthropic's family of models, when to use which one, and make API requests to Claude
- Use multi-modal prompts that combine text and image content blocks, and also work with streaming responses
- Improve your prompting by using prompt templates, using XML to structure prompts, and providing examples
- Implement prompt caching to reduce cost and latency
- Apply tool-use to build a chatbot that can call different tools to respond to queries
- See all these building blocks come together in Computer Use demo
Please sign up here: https://deeplearning.ai/short-courses/building-towards-computer-use-with-anthropic
·225 Views
·0 Reviews