add evox support
This commit is contained in:
@@ -1,21 +0,0 @@
|
||||
import jax, jax.numpy as jnp
|
||||
|
||||
from tensorneat.algorithm import NEAT
|
||||
from tensorneat.genome import DefaultGenome, RecurrentGenome
|
||||
|
||||
key = jax.random.key(0)
|
||||
genome = DefaultGenome(num_inputs=5, num_outputs=3, max_nodes=100, max_conns=500, init_hidden_layers=(1, 2 ,3))
|
||||
state = genome.setup()
|
||||
nodes, conns = genome.initialize(state, key)
|
||||
print(genome.repr(state, nodes, conns))
|
||||
|
||||
inputs = jnp.array([1, 2, 3, 4, 5])
|
||||
transformed = genome.transform(state, nodes, conns)
|
||||
outputs = genome.forward(state, transformed, inputs)
|
||||
|
||||
print(outputs)
|
||||
|
||||
network = genome.network_dict(state, nodes, conns)
|
||||
print(network)
|
||||
|
||||
genome.visualize(network)
|
||||
@@ -1,39 +0,0 @@
|
||||
import jax, jax.numpy as jnp
|
||||
|
||||
from tensorneat.pipeline import Pipeline
|
||||
from tensorneat.algorithm.neat import NEAT
|
||||
from tensorneat.genome import DefaultGenome, DefaultNode, DefaultMutation, BiasNode
|
||||
from tensorneat.problem.func_fit import CustomFuncFit
|
||||
from tensorneat.common import Act, Agg
|
||||
|
||||
|
||||
def pagie_polynomial(inputs):
|
||||
x, y = inputs
|
||||
return x + y
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
genome=DefaultGenome(
|
||||
num_inputs=2,
|
||||
num_outputs=1,
|
||||
max_nodes=3,
|
||||
max_conns=2,
|
||||
init_hidden_layers=(),
|
||||
node_gene=BiasNode(
|
||||
activation_options=[Act.identity],
|
||||
aggregation_options=[Agg.sum],
|
||||
),
|
||||
output_transform=Act.identity,
|
||||
mutation=DefaultMutation(
|
||||
node_add=0,
|
||||
node_delete=0,
|
||||
conn_add=0.0,
|
||||
conn_delete=0.0,
|
||||
)
|
||||
)
|
||||
randkey = jax.random.PRNGKey(42)
|
||||
state = genome.setup()
|
||||
nodes, conns = genome.initialize(state, randkey)
|
||||
print(genome)
|
||||
|
||||
|
||||
34
examples/with_evox/evox_algorithm_adaptor.py
Normal file
34
examples/with_evox/evox_algorithm_adaptor.py
Normal file
@@ -0,0 +1,34 @@
|
||||
import jax.numpy as jnp
|
||||
|
||||
from evox import Algorithm as EvoXAlgorithm, State as EvoXState, jit_class
|
||||
|
||||
from tensorneat.algorithm import BaseAlgorithm as TensorNEATAlgorithm
|
||||
from tensorneat.common import State as TensorNEATState
|
||||
|
||||
|
||||
@jit_class
|
||||
class EvoXAlgorithmAdaptor(EvoXAlgorithm):
|
||||
def __init__(self, algorithm: TensorNEATAlgorithm):
|
||||
self.algorithm = algorithm
|
||||
self.fixed_state = None
|
||||
|
||||
def setup(self, key):
|
||||
neat_algorithm_state = TensorNEATState(randkey=key)
|
||||
neat_algorithm_state = self.algorithm.setup(neat_algorithm_state)
|
||||
self.fixed_state = neat_algorithm_state
|
||||
return EvoXState(alg_state=neat_algorithm_state)
|
||||
|
||||
def ask(self, state: EvoXState):
|
||||
population = self.algorithm.ask(state.alg_state)
|
||||
return population, state
|
||||
|
||||
def tell(self, state: EvoXState, fitness):
|
||||
fitness = jnp.where(jnp.isnan(fitness), -jnp.inf, fitness)
|
||||
neat_algorithm_state = self.algorithm.tell(state.alg_state, fitness)
|
||||
return state.replace(alg_state=neat_algorithm_state)
|
||||
|
||||
def transform(self, individual):
|
||||
return self.algorithm.transform(self.fixed_state, individual)
|
||||
|
||||
def forward(self, transformed, inputs):
|
||||
return self.algorithm.forward(self.fixed_state, transformed, inputs)
|
||||
65
examples/with_evox/example.py
Normal file
65
examples/with_evox/example.py
Normal file
@@ -0,0 +1,65 @@
|
||||
import jax
|
||||
import jax.numpy as jnp
|
||||
|
||||
from evox import workflows, algorithms, problems
|
||||
|
||||
from tensorneat.examples.with_evox.evox_algorithm_adaptor import EvoXAlgorithmAdaptor
|
||||
from tensorneat.examples.with_evox.tensorneat_monitor import TensorNEATMonitor
|
||||
from tensorneat.algorithm import NEAT
|
||||
from tensorneat.algorithm.neat import DefaultSpecies, DefaultGenome, DefaultNodeGene
|
||||
from tensorneat.common import Act
|
||||
|
||||
neat_algorithm = NEAT(
|
||||
species=DefaultSpecies(
|
||||
genome=DefaultGenome(
|
||||
num_inputs=17,
|
||||
num_outputs=6,
|
||||
max_nodes=200,
|
||||
max_conns=500,
|
||||
node_gene=DefaultNodeGene(
|
||||
activation_options=(Act.standard_tanh,),
|
||||
activation_default=Act.standard_tanh,
|
||||
),
|
||||
output_transform=Act.tanh,
|
||||
),
|
||||
pop_size=10000,
|
||||
species_size=10,
|
||||
),
|
||||
)
|
||||
evox_algorithm = EvoXAlgorithmAdaptor(neat_algorithm)
|
||||
|
||||
key = jax.random.PRNGKey(42)
|
||||
model_key, workflow_key = jax.random.split(key)
|
||||
|
||||
monitor = TensorNEATMonitor(neat_algorithm, is_save=False)
|
||||
problem = problems.neuroevolution.Brax(
|
||||
env_name="walker2d",
|
||||
policy=evox_algorithm.forward,
|
||||
max_episode_length=1000,
|
||||
num_episodes=1,
|
||||
backend="mjx"
|
||||
)
|
||||
|
||||
def nan2inf(x):
|
||||
return jnp.where(jnp.isnan(x), -jnp.inf, x)
|
||||
|
||||
# create a workflow
|
||||
workflow = workflows.StdWorkflow(
|
||||
algorithm=evox_algorithm,
|
||||
problem=problem,
|
||||
candidate_transforms=[jax.jit(jax.vmap(evox_algorithm.transform))],
|
||||
fitness_transforms=[nan2inf],
|
||||
monitors=[monitor],
|
||||
opt_direction="max",
|
||||
)
|
||||
|
||||
# init the workflow
|
||||
state = workflow.init(workflow_key)
|
||||
# state = workflow.enable_multi_devices(state)
|
||||
# run the workflow for 100 steps
|
||||
import time
|
||||
|
||||
for i in range(100):
|
||||
tic = time.time()
|
||||
train_info, state = workflow.step(state)
|
||||
monitor.show()
|
||||
@@ -1,6 +0,0 @@
|
||||
import ray
|
||||
|
||||
ray.init(num_gpus=2)
|
||||
|
||||
available_resources = ray.available_resources()
|
||||
print("Available resources:", available_resources)
|
||||
133
examples/with_evox/tensorneat_monitor.py
Normal file
133
examples/with_evox/tensorneat_monitor.py
Normal file
@@ -0,0 +1,133 @@
|
||||
import warnings
|
||||
import os
|
||||
import time
|
||||
import numpy as np
|
||||
|
||||
import jax
|
||||
from jax.experimental import io_callback
|
||||
from evox import Monitor
|
||||
from evox import State as EvoXState
|
||||
|
||||
from tensorneat.algorithm import BaseAlgorithm as TensorNEATAlgorithm
|
||||
from tensorneat.common import State as TensorNEATState
|
||||
|
||||
|
||||
class TensorNEATMonitor(Monitor):
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
neat_algorithm: TensorNEATAlgorithm,
|
||||
save_dir: str = None,
|
||||
is_save: bool = False,
|
||||
):
|
||||
super().__init__()
|
||||
self.neat_algorithm = neat_algorithm
|
||||
|
||||
self.generation_timestamp = time.time()
|
||||
self.alg_state: TensorNEATState = None
|
||||
self.fitness = None
|
||||
self.best_fitness = -np.inf
|
||||
self.best_genome = None
|
||||
|
||||
self.is_save = is_save
|
||||
|
||||
if is_save:
|
||||
if save_dir is None:
|
||||
now = datetime.datetime.now().strftime("%Y%m%d%H%M%S")
|
||||
self.save_dir = f"./{self.__class__.__name__} {now}"
|
||||
else:
|
||||
self.save_dir = save_dir
|
||||
print(f"save to {self.save_dir}")
|
||||
if not os.path.exists(self.save_dir):
|
||||
os.makedirs(self.save_dir)
|
||||
self.genome_dir = os.path.join(self.save_dir, "genomes")
|
||||
if not os.path.exists(self.genome_dir):
|
||||
os.makedirs(self.genome_dir)
|
||||
|
||||
def hooks(self):
|
||||
return ["pre_tell"]
|
||||
|
||||
def pre_tell(self, state: EvoXState, cand_sol, transformed_cand_sol, fitness, transformed_fitness):
|
||||
io_callback(
|
||||
self.store_info,
|
||||
None,
|
||||
state,
|
||||
transformed_fitness,
|
||||
)
|
||||
|
||||
def store_info(self, state: EvoXState, fitness):
|
||||
self.alg_state: TensorNEATState = state.query_state("algorithm").alg_state
|
||||
self.fitness = jax.device_get(fitness)
|
||||
|
||||
def show(self):
|
||||
pop = self.neat_algorithm.ask(self.alg_state)
|
||||
valid_fitnesses = self.fitness[~np.isinf(self.fitness)]
|
||||
|
||||
max_f, min_f, mean_f, std_f = (
|
||||
max(valid_fitnesses),
|
||||
min(valid_fitnesses),
|
||||
np.mean(valid_fitnesses),
|
||||
np.std(valid_fitnesses),
|
||||
)
|
||||
|
||||
new_timestamp = time.time()
|
||||
|
||||
cost_time = new_timestamp - self.generation_timestamp
|
||||
self.generation_timestamp = new_timestamp
|
||||
|
||||
max_idx = np.argmax(self.fitness)
|
||||
if self.fitness[max_idx] > self.best_fitness:
|
||||
self.best_fitness = self.fitness[max_idx]
|
||||
self.best_genome = pop[0][max_idx], pop[1][max_idx]
|
||||
|
||||
if self.is_save:
|
||||
best_genome = jax.device_get((pop[0][max_idx], pop[1][max_idx]))
|
||||
with open(
|
||||
os.path.join(
|
||||
self.genome_dir,
|
||||
f"{int(self.neat_algorithm.generation(self.alg_state))}.npz",
|
||||
),
|
||||
"wb",
|
||||
) as f:
|
||||
np.savez(
|
||||
f,
|
||||
nodes=best_genome[0],
|
||||
conns=best_genome[1],
|
||||
fitness=self.best_fitness,
|
||||
)
|
||||
|
||||
# save best if save path is not None
|
||||
member_count = jax.device_get(self.neat_algorithm.member_count(self.alg_state))
|
||||
species_sizes = [int(i) for i in member_count if i > 0]
|
||||
|
||||
pop = jax.device_get(pop)
|
||||
pop_nodes, pop_conns = pop # (P, N, NL), (P, C, CL)
|
||||
nodes_cnt = (~np.isnan(pop_nodes[:, :, 0])).sum(axis=1) # (P,)
|
||||
conns_cnt = (~np.isnan(pop_conns[:, :, 0])).sum(axis=1) # (P,)
|
||||
|
||||
max_node_cnt, min_node_cnt, mean_node_cnt = (
|
||||
max(nodes_cnt),
|
||||
min(nodes_cnt),
|
||||
np.mean(nodes_cnt),
|
||||
)
|
||||
|
||||
max_conn_cnt, min_conn_cnt, mean_conn_cnt = (
|
||||
max(conns_cnt),
|
||||
min(conns_cnt),
|
||||
np.mean(conns_cnt),
|
||||
)
|
||||
|
||||
print(
|
||||
f"Generation: {self.neat_algorithm.generation(self.alg_state)}, Cost time: {cost_time * 1000:.2f}ms\n",
|
||||
f"\tnode counts: max: {max_node_cnt}, min: {min_node_cnt}, mean: {mean_node_cnt:.2f}\n",
|
||||
f"\tconn counts: max: {max_conn_cnt}, min: {min_conn_cnt}, mean: {mean_conn_cnt:.2f}\n",
|
||||
f"\tspecies: {len(species_sizes)}, {species_sizes}\n",
|
||||
f"\tfitness: valid cnt: {len(valid_fitnesses)}, max: {max_f:.4f}, min: {min_f:.4f}, mean: {mean_f:.4f}, std: {std_f:.4f}\n",
|
||||
)
|
||||
|
||||
# append log
|
||||
if self.is_save:
|
||||
with open(os.path.join(self.save_dir, "log.txt"), "a") as f:
|
||||
f.write(
|
||||
f"{self.neat_algorithm.generation(self.alg_state)},{max_f},{min_f},{mean_f},{std_f},{cost_time}\n"
|
||||
)
|
||||
Reference in New Issue
Block a user