class
ChallengeExtends the Benchmark class to run evaluate the current challenge config and submit results to the remote evaluation server.
Methods
- def evaluate(self, agent: agent.Agent, num_episodes: typing.Optional[int] = None) -> typing.Dict[str, float]
- Evaluates the provide agent in the configured environment either locally or remotely and returns the results dictionary with metrics.
- def local_evaluate(self, agent: agent.Agent, num_episodes: typing.Optional[int] = None) -> typing.Dict[str, float]
- Run evaluation of an Agent in the Env locally.
- def remote_evaluate(self, agent: agent.Agent) -> typing.Dict[str, float]
- Run remote evaluation with evalai for the instantiated Agent and Env. Runs remotely through a challenge evaluation server to prevent any potential for biased results. Imports come challenge-specific dependencies.
- def submit(self, agent)
Special methods
- def __init__(self, eval_remote = False)
Method documentation
def habitat. core. challenge. Challenge. evaluate(self,
agent: agent.Agent,
num_episodes: typing.Optional[int] = None) -> typing.Dict[str, float]
Evaluates the provide agent in the configured environment either locally or remotely and returns the results dictionary with metrics.
Parameters | |
---|---|
agent | agent to be evaluated in environment. |
num_episodes | count of number of episodes for which the evaluation should be run. |
Returns | dict containing metrics tracked by environment. |
def habitat. core. challenge. Challenge. local_evaluate(self,
agent: agent.Agent,
num_episodes: typing.Optional[int] = None) -> typing.Dict[str, float]
Run evaluation of an Agent in the Env locally.
Parameters | |
---|---|
agent | The Agent to evaluate. |
num_episodes | The number of episodes to evaluate. |
Returns | The results dictionary containing metrics. |
def habitat. core. challenge. Challenge. remote_evaluate(self,
agent: agent.Agent) -> typing.Dict[str, float]
Run remote evaluation with evalai for the instantiated Agent and Env. Runs remotely through a challenge evaluation server to prevent any potential for biased results. Imports come challenge-specific dependencies.
Parameters | |
---|---|
agent | The Agent to evaluate. |
Returns | The results dictionary containing metrics. |