import collections from functools import partial as bind import elements import embodied import numpy as np def train_eval( make_agent, make_replay_train, make_replay_eval, make_env_train, make_env_eval, make_stream, make_logger, args): agent = make_agent() replay_train = make_replay_train() replay_eval = make_replay_eval() logger = make_logger() logdir = elements.Path(args.logdir) logdir.mkdir() print('Logdir', logdir) step = logger.step usage = elements.Usage(**args.usage) agg = elements.Agg() train_episodes = collections.defaultdict(elements.Agg) train_epstats = elements.Agg() eval_episodes = collections.defaultdict(elements.Agg) eval_epstats = elements.Agg() policy_fps = elements.FPS() train_fps = elements.FPS() batch_steps = args.batch_size * args.batch_length should_train = elements.when.Ratio(args.train_ratio / batch_steps) should_log = elements.when.Clock(args.log_every) should_report = elements.when.Clock(args.report_every) should_save = elements.when.Clock(args.save_every) @elements.timer.section('logfn') def logfn(tran, worker, mode): episodes = dict(train=train_episodes, eval=eval_episodes)[mode] epstats = dict(train=train_epstats, eval=eval_epstats)[mode] episode = episodes[worker] tran['is_first'] and episode.reset() episode.add('score', tran['reward'], agg='sum') episode.add('length', 1, agg='sum') episode.add('rewards', tran['reward'], agg='stack') for key, value in tran.items(): if value.dtype == np.uint8 and value.ndim == 3: if worker == 0: episode.add(f'policy_{key}', value, agg='stack') elif key.startswith('log/'): assert value.ndim == 0, (key, value.shape, value.dtype) episode.add(key + '/avg', value, agg='avg') episode.add(key + '/max', value, agg='max') episode.add(key + '/sum', value, agg='sum') if tran['is_last']: result = episode.result() logger.add({ 'score': result.pop('score'), 'length': result.pop('length'), }, prefix='episode') rew = result.pop('rewards') if len(rew) > 1: result['reward_rate'] = (np.abs(rew[1:] - rew[:-1]) >= 0.01).mean() epstats.add(result) fns = [bind(make_env_train, i) for i in range(args.envs)] driver_train = embodied.Driver(fns, parallel=(not args.debug)) driver_train.on_step(lambda tran, _: step.increment()) driver_train.on_step(lambda tran, _: policy_fps.step()) driver_train.on_step(replay_train.add) driver_train.on_step(bind(logfn, mode='train')) fns = [bind(make_env_eval, i) for i in range(args.eval_envs)] driver_eval = embodied.Driver(fns, parallel=(not args.debug)) driver_eval.on_step(replay_eval.add) driver_eval.on_step(bind(logfn, mode='eval')) driver_eval.on_step(lambda tran, _: policy_fps.step()) stream_train = iter(agent.stream(make_stream(replay_train, 'train'))) stream_report = iter(agent.stream(make_stream(replay_train, 'report'))) stream_eval = iter(agent.stream(make_stream(replay_eval, 'eval'))) carry_train = [agent.init_train(args.batch_size)] carry_report = agent.init_report(args.batch_size) carry_eval = agent.init_report(args.batch_size) def trainfn(tran, worker): if len(replay_train) < args.batch_size * args.batch_length: return for _ in range(should_train(step)): with elements.timer.section('stream_next'): batch = next(stream_train) carry_train[0], outs, mets = agent.train(carry_train[0], batch) train_fps.step(batch_steps) if 'replay' in outs: replay_train.update(outs['replay']) agg.add(mets, prefix='train') driver_train.on_step(trainfn) def reportfn(carry, stream): agg = elements.Agg() for _ in range(args.report_batches): batch = next(stream) carry, mets = agent.report(carry, batch) agg.add(mets) return carry, agg.result() cp = elements.Checkpoint(logdir / 'ckpt') cp.step = step cp.agent = agent cp.replay_train = replay_train cp.replay_eval = replay_eval if args.from_checkpoint: elements.checkpoint.load(args.from_checkpoint, dict( agent=bind(agent.load, regex=args.from_checkpoint_regex))) cp.load_or_save() should_save(step) # Register that we just saved. print('Start training loop') train_policy = lambda *args: agent.policy(*args, mode='train') eval_policy = lambda *args: agent.policy(*args, mode='eval') driver_train.reset(agent.init_policy) while step < args.steps: if should_report(step): print('Evaluation') driver_eval.reset(agent.init_policy) driver_eval(eval_policy, episodes=args.eval_eps) logger.add(eval_epstats.result(), prefix='epstats') if len(replay_train): carry_report, mets = reportfn(carry_report, stream_report) logger.add(mets, prefix='report') if len(replay_eval): carry_eval, mets = reportfn(carry_eval, stream_eval) logger.add(mets, prefix='eval') driver_train(train_policy, steps=10) if should_log(step): logger.add(agg.result()) logger.add(train_epstats.result(), prefix='epstats') logger.add(replay_train.stats(), prefix='replay') logger.add(usage.stats(), prefix='usage') logger.add({'fps/policy': policy_fps.result()}) logger.add({'fps/train': train_fps.result()}) logger.add({'timer': elements.timer.stats()['summary']}) logger.write() if should_save(step): cp.save() logger.close()