Back in December I gave a talk introducing Numpy to the PyNorthwest group. The slides are available as a pdf.

Although I frequently use Numpy I'm far from an expert and the content of my talk reflected this. I started with a general introduction to the array object and then expanded the scope of the talk to highlight some of the projects that use Numpy. I gave an example of using MDP and matplotlib.

The talk was followed by some excellent discussion. We went through some of the code on slide 6 in a lot of detail.

The PyNorthwest group meets at Madlab in Manchester city centre on the third Thursday of each month. If you're in the area check it out. The January event is on the 19th, starting at 7pm.

I've been using MDP and matplotlib a lot recently and although overall I've been very pleased with the documentation for both projects I have run into a few problems for which the solutions were not immediately obvious. This post gives the solution for each in the expectation it will certainly be useful to me in the future and the hope that it may also be useful to others.

The tutorial for the Modular Toolkit for Data Processing (MDP) starts with a quick example of using the toolkit for a pca analysis and yet I still ran into a couple of problems. The first issue I had was how the pca function expects to receive data. I suspect this is simply due to unfamiliarity with the field and the language used within the field. For future reference the data is expected to be in the following format.

Gene 1 | Gene 2 | Gene 3 | Gene 4 | |

Experimental Condition 1 | . | . | . | . |

Experimental Condition 2 | . | . | . | . |

The previously mentioned quick start tutorial was very useful in getting something useful out quickly but I couldn't find a way to get a value for how much of the variance present in the data was accounted for in the principal components. To get that, as far as I've been able to determine, you need to interact with the PCANode directly rather than using the convenience function. The code is still relative straightforward.

```
import mdp
import numpy as np
import matplotlib.pyplot as plt
#Create sample data
var1 = np.random.normal(loc=0., scale=0.5, size=(10,5))
var2 = np.random.normal(loc=4., scale=1., size=(10,5))
var = np.concatenate((var1,var2), axis=0)
#Create the PCA node and train it
pcan = mdp.nodes.PCANode(output_dim=3)
pcar = pcan.execute(var)
#Graph the results
fig = plt.figure()
ax = fig.add_subplot(111)
ax.plot(pcar[:10,0], pcar[:10,1], 'bo')
ax.plot(pcar[10:,0], pcar[10:,1], 'ro')
#Show variance accounted for
ax.set_xlabel('PC1 (%.3f%%)' % (pcan.d[0]))
ax.set_ylabel('PC2 (%.3f%%)' % (pcan.d[1]))
plt.show()
```

Running this code produces an image similar to the one below.

The growing neural gas implementation was another sample application highlighted in the tutorial for MDP. It held my interest for a while as a technique which could potentially be applied to the transcription of plaques for the openplaques project. It wasn't immediately obvious how to get the position of a node from a connected nodes object. As the tutorial left the details of visualisation up to the user I'll present the solution to getting the node location in the form of the necessary code to visualise the node training. The end result will look something like the following.

I've been using Matplotlib to plot data exclusively for a while now. The defaults produce reasonable quality graphs and any differences in opinion can be quickly fixed either by altering options in matplotlib or, as the graphs can be saved in svg format, in a vector image manipulation program such as Inkscape. Although most options can be changed in matplotlib it can sometimes be difficult to find the correct option. Most of the time the naming of variables are, to my mind, logical but sometimes I just can't find the right way to describe what I want to do.

I wanted to have a grid of 6 graphs but didn't want to display the axes on all the graphs as I felt this looked cluttered.

If I was going to display the axes on only some of the graphs then the values for the axes needed to be the same on all of them.

```
import numpy as np
import matplotlib.pyplot as plt
#Generate sample data
var = np.random.random_sample((40,2))
fig = plt.figure()
for i in range(4):
ax = fig.add_subplot(220 + i + 1)
start = i * 10
ax.plot(var[start:start+10,0], var[start:start+10,1], 'bo')
#Hide the x axis on the top row of charts
if i in [0,1]:
ax.set_xticklabels(ax.get_xticklabels(), visible=False)
#Hide the y axis on the right column of charts
if i in [1,3]:
ax.set_yticklabels(ax.get_yticklabels(), visible=False)
#Set the axis range
ax.axis([0,1,0,1])
plt.show()
```

Running this code should produce an image similar to the one below.

The legend assumes that values are connected so two points and the connecting line are shown by default. If the points on the graph aren't connected then this looked strange. To remove the duplicate symbol is straightforward.

```
import numpy as np
import matplotlib.pyplot as plt
#Generate sample data
var = np.random.random_sample((10,2))
#Plot data with labels
fig = plt.figure()
ax = fig.add_subplot(111)
ax.plot(var[0:5,0], var[0:5,1], 'bo', label="First half")
ax.plot(var[5:10,0], var[5:10,1], 'r^', label="Second half")
ax.legend(numpoints=1)
plt.show()
```

Page 1 of 1

- Lightning talk slides on deep learning with keras
- Lightning talk slides on web server log analysis with pandas
- Seizure detection challenge on kaggle
- Introduction to Scientific Computing at DC Python
- Strategies for data collection
- Disk space used by databases and tables in MySQL PostgreSQL
- Django and Scrapy
- Numpy talk at Python Northwest
- Full text visualisation
- Images and Vision in Python: Slides from talk at Python Edinburgh Mini-Conf 2011

- 3D printing
- aicookbook
- Alectruino
- amazon
- Arduino
- backups
- barcamp
- BarcampNE
- BarcampNE3
- BarcampNE4
- barcampnortheast3
- bcNE3
- bcNE4
- Clock
- CNC
- DIYBio
- django
- Electronics
- enchant
- hiv
- jungledisk
- kaggle
- keras
- linode
- lxml
- Maker faire
- matplotlib
- mdp
- mfuk
- MySQL
- Newcastle Science Festival
- ocr
- openplaques
- pandas
- pdfminer
- PHP
- phpass
- PIL
- Postgresql
- prediction
- pynw
- python
- s3
- scikit-learn
- scrapy
- security
- tarsnap
- tesseract
- vps
- Zend_Auth
- zendframework
- Zend_HTTP
- Zend_Mail
- Zend_View