Spaces:
Running
Running
# Features and Options | |
You likely don't need to tune the hyperparameters yourself, | |
but if you would like, you can use `hyperparamopt.py` as an example. | |
Some configurable features and options in `PySR` which you | |
may find useful include: | |
- `binary_operators`, `unary_operators` | |
- `niterations` | |
- `ncyclesperiteration` | |
- `procs` | |
- `populations` | |
- `weights` | |
- `maxsize`, `maxdepth` | |
- `batching`, `batchSize` | |
- `variable_names` (or pandas input) | |
- Constraining operator complexity | |
- LaTeX, SymPy | |
- Callable exports: numpy, pytorch, jax | |
- `loss` | |
These are described below | |
The program will output a pandas DataFrame containing the equations, | |
mean square error, and complexity. It will also dump to a csv | |
at the end of every iteration, | |
which is `hall_of_fame_{date_time}.csv` by default. It also prints the | |
equations to stdout. | |
## Operators | |
A list of operators can be found on the operators page. | |
One can define custom operators in Julia by passing a string: | |
```python | |
equations = pysr.pysr(X, y, niterations=100, | |
binary_operators=["mult", "plus", "special(x, y) = x^2 + y"], | |
extra_sympy_mappings={'special': lambda x, y: x**2 + y}, | |
unary_operators=["cos"]) | |
``` | |
Now, the symbolic regression code can search using this `special` function | |
that squares its left argument and adds it to its right. Make sure | |
all passed functions are valid Julia code, and take one (unary) | |
or two (binary) float32 scalars as input, and output a float32. This means if you | |
write any real constants in your operator, like `2.5`, you have to write them | |
instead as `2.5f0`, which defines it as `Float32`. | |
Operators are automatically vectorized. | |
One should also define `extra_sympy_mappings`, | |
so that the SymPy code can understand the output equation from Julia, | |
when constructing a useable function. This step is optional, but | |
is necessary for the `lambda_format` to work. | |
One can also edit `operators.jl`. | |
## Iterations | |
This is the total number of generations that `pysr` will run for. | |
I usually set this to a large number, and exit when I am satisfied | |
with the equations. | |
## Cycles per iteration | |
Each cycle considers every 10-equation subsample (re-sampled for each individual 10, | |
unless `fast_cycle` is set in which case the subsamples are separate groups of equations) | |
a single time, producing one mutated equation for each. | |
The parameter `ncyclesperiteration` defines how many times this | |
occurs before the equations are compared to the hall of fame, | |
and new equations are migrated from the hall of fame, or from other populations. | |
It also controls how slowly annealing occurs. You may find that increasing | |
`ncyclesperiteration` results in a higher cycles-per-second, as the head | |
worker needs to reduce and distribute new equations less often, and also increases | |
diversity. But at the same | |
time, a smaller number it might be that migrating equations from the hall of fame helps | |
each population stay closer to the best current equations. | |
## Processors | |
One can adjust the number of workers used by Julia with the | |
`procs` option. You should set this equal to the number of cores | |
you want `pysr` to use. This will also run `procs` number of | |
populations simultaneously by default. | |
## Populations | |
By default, `populations=procs`, but you can set a different | |
number of populations with this option. More populations may increase | |
the diversity of equations discovered, though will take longer to train. | |
However, it may be more efficient to have `populations>procs`, | |
as there are multiple populations running | |
on each core. | |
## Weighted data | |
Here, we assign weights to each row of data | |
using inverse uncertainty squared. We also use 10 processes | |
instead of the usual 4, which creates more populations | |
(one population per thread). | |
```python | |
sigma = ... | |
weights = 1/sigma**2 | |
equations = pysr.pysr(X, y, weights=weights, procs=10) | |
``` | |
## Max size | |
`maxsize` controls the maximum size of equation (number of operators, | |
constants, variables). `maxdepth` is by default not used, but can be set | |
to control the maximum depth of an equation. These will make processing | |
faster, as longer equations take longer to test. | |
One can warm up the maxsize from a small number to encourage | |
PySR to start simple, by using the `warmupMaxsize` argument. | |
This specifies that maxsize increases every `warmupMaxsize`. | |
## Batching | |
One can turn on mini-batching, with the `batching` flag, | |
and control the batch size with `batchSize`. This will make | |
evolution faster for large datasets. Equations are still evaluated | |
on the entire dataset at the end of each iteration to compare to the hall | |
of fame, but only on a random subset during mutations and annealing. | |
## Variable Names | |
You can pass a list of strings naming each column of `X` with | |
`variable_names`. Alternatively, you can pass `X` as a pandas dataframe | |
and the columns will be used as variable names. Make sure only | |
alphabetical characters and `_` are used in these names. | |
## Constraining operator complexity | |
One can limit the complexity of specific operators with the `constraints` parameter. | |
There is a "maxsize" parameter to PySR, but there is also an operator-level | |
"constraints" parameter. One supplies a dict, like so: | |
```python | |
constraints={'pow': (-1, 1), 'mult': (3, 3), 'cos': 5} | |
``` | |
What this says is that: a power law x^y can have an expression of arbitrary (-1) complexity in the x, but only complexity 1 (e.g., a constant or variable) in the y. So (x0 + 3)^5.5 is allowed, but 5.5^(x0 + 3) is not. | |
I find this helps a lot for getting more interpretable equations. | |
The other terms say that each multiplication can only have sub-expressions | |
of up to complexity 3 (e.g., 5.0 + x2) in each side, and cosine can only operate on | |
expressions of complexity 5 (e.g., 5.0 + x2 exp(x3)). | |
## LaTeX, SymPy | |
The `pysr` command will return a pandas dataframe. The `sympy_format` | |
column gives sympy equations, and the `lambda_format` gives callable | |
functions. These use the variable names you have provided. | |
There are also some helper functions for doing this quickly. | |
You can call `get_hof()` (or pass an equation file explicitly to this) | |
to get this pandas dataframe. | |
You can call the functions `best()` to get the sympy format | |
for the best equation, using the `score` column to sort equations. | |
`best_latex()` returns the LaTeX form of this, and `best_callable()` | |
returns a callable function. | |
## Callable exports: numpy, pytorch, jax | |
By default, the dataframe of equations will contain columns | |
with the identifier `lambda_format`. These are simple functions | |
which correspond to the equation, but executed | |
with numpy functions. You can pass your `X` matrix to these functions | |
just as you did to the `pysr` call. Thus, this allows | |
you to numerically evaluate the equations over different output. | |
One can do the same thing for PyTorch, which uses code | |
from [sympytorch](https://github.com/patrick-kidger/sympytorch), | |
and for JAX, which uses code from | |
[sympy2jax](https://github.com/MilesCranmer/sympy2jax). | |
For torch, set the argument `output_torch_format=True`, which | |
will generate a column `torch_format`. Each element of this column | |
is a PyTorch module which runs the equation, using PyTorch functions, | |
over `X` (as a PyTorch tensor). This is differentiable, and the | |
parameters of this PyTorch module correspond to the learned parameters | |
in the equation, and are trainable. | |
For jax, set the argument `output_jax_format=True`, which | |
will generate a column `jax_format`. Each element of this column | |
is a dictionary containing a `'callable'` (a JAX function), | |
and `'parameters'` (a list of parameters in the equation). | |
One can execute this function with: `element['callable'](X, element['parameters'])`. | |
Since the parameter list is a jax array, this therefore lets you also | |
train the parameters within JAX (and is differentiable). | |
If you forget to turn these on when calling the function initially, | |
you can re-run `get_hof(output_jax_format=True)`, and it will re-use | |
the equations and other state properties, assuming you haven't | |
re-run `pysr` in the meantime! | |
## `loss` | |
The default loss is mean-square error, and weighted mean-square error. | |
One can pass an arbitrary Julia string to define a custom loss, using, | |
e.g., `loss="myloss(x, y) = abs(x - y)^1.5"`. For more details, | |
see the | |
[Losses](https://milescranmer.github.io/SymbolicRegression.jl/dev/losses/) | |
page for SymbolicRegression.jl. | |
Here are some additional examples: | |
abs(x-y) loss | |
```python | |
pysr(..., loss="f(x, y) = abs(x - y)^1.5") | |
``` | |
Note that the function name doesn't matter: | |
```python | |
pysr(..., loss="loss(x, y) = abs(x * y)") | |
``` | |
With weights: | |
```python | |
pysr(..., weights=weights, loss="myloss(x, y, w) = w * abs(x - y)") | |
``` | |
Weights can be used in arbitrary ways: | |
```python | |
pysr(..., weights=weights, loss="myloss(x, y, w) = abs(x - y)^2/w^2") | |
``` | |
Built-in loss (faster) (see [losses](https://astroautomata.com/SymbolicRegression.jl/dev/losses/)). | |
This one computes the L3 norm: | |
```python | |
pysr(..., loss="LPDistLoss{3}()") | |
``` | |
Can also uses these losses for weighted (weighted-average): | |
```python | |
pysr(..., weights=weights, loss="LPDistLoss{3}()") | |
``` | |