You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Talk about my project that will help users to evaluate Agents against Environment at Scale.
Duration
30 min
Outline
The rise of reinforcement learning based problems or any problem which requires that an agent must interact with an environment introduces additional challenges for benchmarking. In contrast to the supervised learning setting where performance is measured by evaluating on a static test set, it is less straightforward to measure generalization performance of these agents in the context of the interactions with the environment. Evaluating these agents involves running the associated code on a collection of unseen environments that constitutes a hidden test set for such a scenario. This project deals with seting up a robust pipeline for uploading prediction code in the form of Docker containers (as opposed to test prediction file) that will be evaluated on remote machines.
Evaluating AI Agents against Environment at Scale
Description
Talk about my project that will help users to evaluate Agents against Environment at Scale.
Duration
Outline
Additional notes
The text was updated successfully, but these errors were encountered: