# 6. Use examples¶

This page is still under development, so most of the headers are just place-holders for the actual examples

Note

The examples provided here are not meant as a through description of AequilibraE’s capabilities. For that, please look into the API documentation or email aequilibrae@googlegroups.com

## 6.1. Logging¶

AequilibraE uses Python’s standard logging library to a file called
*aequilibrae.log*, but the output folder for this logging can be changed to a
custom system folder by altering the parameter **system –> logging_directory** on
the parameters file.

As an example, one could do programatically change the output folder to
*‘D:/myProject/logs’* by doing the following:

```
from aequilibrae import Parameters
fldr = 'D:/myProject/logs'
p = Parameters()
p.parameters['system']['logging_directory'] = fldr
p.writeback()
```

The other useful resource, especially during model debugging it to also show all log messages directly on the screen. Doing that requires a little knowledge of the Python Logging library, but it is just as easy:

```
from aequilibrae import logger
import logging
stdout_handler = logging.StreamHandler(sys.stdout)
logger.addHandler(stdout_handler)
```

## 6.2. Parameters module¶

Several examples on how to manipulate the parameters within AequilibraE have been shown in other parts of this tutorial.

However, in case you ever have trouble with parameter changes you have made,
you can always revert them to their default values. But remember, **ALL**
**CHANGES WILL BE LOST**.

```
from aequilibrae import Parameters
fldr = 'D:/myProject/logs'
p = Parameters()
p.reset_default()
```

## 6.3. Project module¶

Let’s suppose one wants to create project files for a list of 20 cities around the world with their complete networks downloaded from Open Street Maps and place them on a local folder for analysis at a later time.

```
from aequilibrae.project import Project
import os
cities = ["Darwin, Australia",
"Karlsruhe, Germany",
"London, UK",
"Paris, France",
"Shanghai, China",
"Sao Paulo, Brazil",
"Rio de Janeiro, Brazil",
"Los Angeles, USA",
"New York, USA",
"Mexico City, Mexico",
"Berlin, Germany",
"Vancouver, Canada",
"Montreal, Canada",
"Toronto, Canada",
"Madrid, Spain",
"Lisbon, Portugal",
"Rome, Italy",
"Perth, Australia",
"Hobart, Australia",
"Auckland, New Zealand"]
for city in cities:
pth = f'd:/net_tests/{city}.sqlite'
p = Project(pth, True)
p.network.create_from_osm(place_name=city)
p.conn.close()
del p
```

## 6.4. Paths module¶

```
from aequilibrae.paths import allOrNothing
from aequilibrae.paths import path_computation
from aequilibrae.paths.results import AssignmentResults as asgr
from aequilibrae.paths.results import PathResults as pthr
```

### 6.4.1. Path computation¶

### 6.4.2. Skimming¶

Let’s suppose you want to compute travel times between all zone on your network. In that case, you need only a graph that you have previously built, and the list of skims you want to compute.

```
from aequilibrae.paths.results import SkimResults as skmr
from aequilibrae.paths import Graph
from aequilibrae.paths import NetworkSkimming
# We instantiate the graph and load it from disk (say you created it using the QGIS GUI
g = Graph()
g.load_from_disk(aeg_pth)
# You now have to set the graph for what you want
# In this case, we are computing fastest path (minimizing free flow time)
# We are also **blocking** paths from going through centroids
g.set_graph(cost_field='fftime', block_centroid_flows=True)
# We will be skimming for fftime **AND** length along the way
g.set_skimming(['fftime', 'length'])
# We instantiate the skim results and prepare it to have results compatible with the graph provided
result = skmr()
result.prepare(g)
# We create the network skimming object and execute it
# This is multi-threaded, so if the network is too big, prepare for a slow computer
skm = NetworkSkimming(g, result)
skm.execute()
```

If you want to use fewer cores for this computation (which also saves memory), you also can do it
You just need to use the method *set_cores* before you run the skimming. Ideally it is done before preparing it

```
result = skmr()
result.set_cores(3)
result.prepare(g)
```

And if you want to compute skims between all nodes in the network, all you need to do is to make sure the list of centroids in your graph is updated to include all nodes in the graph

```
from aequilibrae.paths.results import SkimResults as skmr
from aequilibrae.paths import Graph
from aequilibrae.paths import NetworkSkimming
g = Graph()
g.load_from_disk(aeg_pth)
# Let's keep the original list of centroids in case we want to use it again
orig_centr = g.centroids
# Now we set the list of centroids to include all nodes in the network
g.prepare_graph(g.all_nodes)
# And continue **almost** like we did before
# We just need to remember to NOT block paths through centroids. Otherwise there will be no paths available
g.set_graph(cost_field='fftime', block_centroid_flows=False)
g.set_skimming('fftime')
result = skmr()
result.prepare(g)
skm = NetworkSkimming(g, result)
skm.execute()
```

Setting skimming after setting the graph is **CRITICAL**, and the skim matrices are part of the result object.

You can save the results to your place of choice in AequilibraE format or export to OMX or CSV

```
result.skims.export('path/to/desired/folder/file_name.omx')
result.skims.export('path/to/desired/folder/file_name.csv')
result.skims.copy('path/to/desired/folder/file_name.aem')
```

### 6.4.3. Traffic Assignment¶

```
some code
```

### 6.4.4. Advanced usage: Building a Graph¶

Let’s suppose now that you are interested in creating links from a bespoke procedure. For the purpose of this example, let’s say you have a sparse matrix representing a graph as an adjacency matrix

```
from aequilibrae.paths import Graph
from aequilibrae import reserved_fields
from scipy.sparse import coo_matrix
# original_adjacency_matrix is a sparse matrix where positive values are actual links
# where the value of the cell is the distance in that link
# We create the sparse matrix in proper sparse matrix format
sparse_graph = coo_matrix(original_adjacency_matrix)
# We create the structure to create the network
all_types = [k._Graph__integer_type,
k._Graph__integer_type,
k._Graph__integer_type,
np.int8,
k._Graph__float_type,
k._Graph__float_type]
all_titles = [reserved_fields.link_id,
reserved_fields.a_node,
reserved_fields.b_node,
reserved_fields.direction,
"length_ab",
"length_ba"]
dt = [(t, d) for t, d in zip(all_titles, all_types)]
# Number of links
num_links = sparse_graph.data.shape[0]
my_graph = Graph()
my_graph.network = np.zeros(links, dtype=dt)
my_graph.network[reserved_fields.link_id] = np.arange(links) + 1
my_graph.network[reserved_fields.a_node] = sparse_graph.row
my_graph.network[reserved_fields.b_node] = sparse_graph.col
my_graph.network["length_ab"] = sparse_graph.data
# If the links are directed (from A to B), direction is 1. If bi-directional, use zeros
my_graph.network[reserved_fields.direction] = np.ones(links)
# If uni-directional from A to B the value is not used
my_graph.network["length_ba"] = mat.data * 10000
# Let's say that all nodes in the network are centroids
list_of_centroids = np.arange(max(sparse_graph.shape[0], sparse_graph.shape[0])+ 1)
centroids_list = np.array(list_of_centroids)
my_graph.type_loaded = 'NETWORK'
my_graph.status = 'OK'
my_graph.network_ok = True
my_graph.prepare_graph(centroids_list)
```

This usage is really advanced, and very rarely not-necessary. Make sure to know what you are doing before going down this route

## 6.5. Trip distribution¶

The support for trip distribution in AequilibraE is not very comprehensive, mostly because of the loss of relevance that such type of model has suffered in the last decade.

However, it is possible to calibrate and apply synthetic gravity models and to perform Iterative Proportional Fitting (IPF) with really high performance, which might be of use in many applications other than traditional distribution.

### 6.5.1. Synthetic gravity calibration¶

```
some code
```

### 6.5.2. Synthetic gravity application¶

In this example, imagine that you have your demographic information in an sqlite database and that you have already computed your skim matrix.

It is also important to notice that it is crucial to have consistent data, such as same set of zones (indices) in both the demographics and the impedance matrix.

```
import pandas as pd
import sqlite3
from aequilibrae.matrix import AequilibraeMatrix
from aequilibrae.matrix import AequilibraeData
from aequilibrae.distribution import SyntheticGravityModel
from aequilibrae.distribution import GravityApplication
# We define the model we will use
model = SyntheticGravityModel()
# Before adding a parameter to the model, you need to define the model functional form
model.function = "GAMMA" # "EXPO" or "POWER"
# Only the parameter(s) applicable to the chosen functional form will have any effect
model.alpha = 0.1
model.beta = 0.0001
# Or you can load the model from a file
model.load('path/to/model/file')
# We load the impedance matrix
matrix = AequilibraeMatrix()
matrix.load('path/to/impedance_matrix.aem')
matrix.computational_view(['distance'])
# We create the vectors we will use
conn = sqlite3.connect('path/to/demographics/database')
query = "SELECT zone_id, population, employment FROM demographics;"
df = pd.read_sql_query(query,conn)
index = df.zone_id.values[:]
zones = index.shape[0]
# You create the vectors you would have
df = df.assign(production=df.population * 3.0)
df = df.assign(attraction=df.employment * 4.0)
# We create the vector database
args = {"entries": zones, "field_names": ["productions", "attractions"],
"data_types": [np.float64, np.float64], "memory_mode": True}
vectors = AequilibraeData()
vectors.create_empty(**args)
# Assign the data to the vector object
vectors.productions[:] = df.production.values[:]
vectors.attractions[:] = df.attraction.values[:]
vectors.index[:] = zones[:]
# Balance the vectors
vectors.attractions[:] *= vectors.productions.sum() / vectors.attractions.sum()
args = {"impedance": matrix,
"rows": vectors,
"row_field": "productions",
"model": model,
"columns": vectors,
"column_field": "attractions",
"output": 'path/to/output/matrix.aem',
"nan_as_zero":True
}
gravity = GravityApplication(**args)
gravity.apply()
```

### 6.5.3. Iterative Proportional Fitting (IPF)¶

The implementation of IPF is fully vectorized and leverages all the speed of NumPy, but it does not include the fancy multithreading implemented in path computation.

**Please note that the AequilibraE matrix used as input is OVERWRITTEN by the IPF**

```
import pandas as pd
from aequilibrae.distribution import Ipf
from aequilibrae.matrix import AequilibraeMatrix
from aequilibrae.matrix import AequilibraeData
matrix = AequilibraeMatrix()
# Here we can create from OMX or load from an AequilibraE matrix.
matrix.create_from_omx(path/to/aequilibrae_matrix, path/to/omxfile)
# The matrix will be operated one (see the note on overwriting), so it does
# not make sense load an OMX matrix
source_vectors = pd.read_csv(path/to/CSVs)
zones = source_vectors.zone.shape[0]
args = {"entries": zones, "field_names": ["productions", "attractions"],
"data_types": [np.float64, np.float64], "memory_mode": True}
vectors = AequilibraEData()
vectors.create_empty(**args)
vectors.productions[:] = source_vectors.productions[:]
vectors.attractions[:] = source_vectors.attractions[:]
# We assume that the indices would be sorted and that they would match the matrix indices
vectors.index[:] = source_vectors.zones[:]
args = {
"matrix": matrix, "rows": vectors, "row_field": "productions", "columns": vectors,
"column_field": "attractions", "nan_as_zero": False}
fratar = Ipf(**args)
fratar.fit()
# We can get back to our OMX matrix in the end
matrix.export(path/to_omx/output)
```

## 6.6. Transit¶

We only have import for now, and it is likely to not work on Windows if you want the geometries

### 6.6.1. GTFS import¶

```
some code
```

## 6.7. Matrices¶

Lets say we want to Import the freight matrices provided with FAF into AequilibraE’s matrix format in order to create some Delaunay Lines in QGIS or to perform traffic assignment

### 6.7.1. Required data¶

### 6.7.2. Useful Information¶

### 6.7.3. The code¶

We import all libraries we will need, including the AequilibraE

```
import pandas as pd
import numpy as np
import os
from aequilibrae.matrix import AequilibraeMatrix
from scipy.sparse import coo_matrix
```

Now we set all the paths for files and parameters we need and import the matrices into a Pandas DataFrame

```
data_folder = 'Y:/ALL DATA/DATA/Pedro/Professional/Data/USA/FAF/4.4'
data_file = 'FAF4.4_HiLoForecasts.csv'
sctg_names_file = 'sctg_codes.csv' # Simplified to 50 characters, which is AequilibraE's limit
output_folder = data_folder
matrices = pd.read_csv(os.path.join(data_folder, data_file), low_memory=False)
```

We import the sctg codes

```
sctg_names = pd.read_csv(os.path.join(data_folder, sctg_names_file), low_memory=False)
sctg_names.set_index('Code', inplace=True)
sctg_descr = list(sctg_names['Commodity Description'])
```

We now process the matrices to collect all the data we need, such as:

List of zones

CSTG codes

Matrices/scenarios we are importing

```
all_zones = np.array(sorted(list(set( list(matrices.dms_orig.unique()) + list(matrices.dms_dest.unique())))))
# Count them and create a 0-based index
num_zones = all_zones.shape[0]
idx = np.arange(num_zones)
# Creates the indexing dataframes
origs = pd.DataFrame({"from_index": all_zones, "from":idx})
dests = pd.DataFrame({"to_index": all_zones, "to":idx})
# adds the new index columns to the pandas dataframe
matrices = matrices.merge(origs, left_on='dms_orig', right_on='from_index', how='left')
matrices = matrices.merge(dests, left_on='dms_dest', right_on='to_index', how='left')
# Lists sctg codes and all the years/scenarios we have matrices for
mat_years = [x for x in matrices.columns if 'tons' in x]
sctg_codes = matrices.sctg2.unique()
```

We now import one matrix for each year, saving all the SCTG codes as different matrix cores in our zoning system

```
# aggregate the matrix according to the relevant criteria
agg_matrix = matrices.groupby(['from', 'to', 'sctg2'])[mat_years].sum()
# returns the indices
agg_matrix.reset_index(inplace=True)
for y in mat_years:
mat = AequilibraeMatrix()
# Here it does not make sense to use OMX
# If one wants to create an OMX from other data sources, openmatrix is
# the library to use
kwargs = {'file_name': os.path.join(output_folder, y + '.aem'),
'zones': num_zones,
'matrix_names': sctg_descr}
mat.create_empty(**kwargs)
mat.index[:] = all_zones[:]
# for all sctg codes
for i in sctg_names.index:
prod_name = sctg_names['Commodity Description'][i]
mat_filtered_sctg = agg_matrix[agg_matrix.sctg2 == i]
m = coo_matrix((mat_filtered_sctg[y], (mat_filtered_sctg['from'], mat_filtered_sctg['to'])),
shape=(num_zones, num_zones)).toarray().astype(np.float64)
mat.matrix[prod_name][:,:] = m[:,:]
mat.close()
```