Django Deployment

I was working on a tool to make django apps really easy to deploy. Just write a config file and run a command specifying the directory of the app. I think it will deserve another post when it’s finished. In the meantime you can take a look at the code on It’s based on the amazing ssh tool for python, fabric.

Any feedback or contribution would be appreciated!


Finally i developed my Ruby on Rails application, A Web Chat for Mathematics! Its’ called MathChat

MathChat supports Latex commands to draw complex mathematical symbols. If you don’t heard about Latex, here is a little tutorial about it.

And here is a little guide to deploy and run your own MathChat server:

In order to install the dependencies of MathChat you must do the follow in your *unix system (tested on ubuntu 10.04 and 10.10):

~$ sudo apt-get install ruby build-essential libopenssl-ruby libsqlite3-dev rubygems ruby1.8-dev
~$ sudo gem install rubygems-update --version=1.3.7
~$ sudo gem install sqlite3-ruby
~$ sudo gem install -v=2.3.5 rails

That setup an environment to run MathChat and give you the posibility to continue developing the application!
If you know about ruby on rails you can modify and expand the application as you want. It’s really easy.

To run the MathChat server in localhost:3000 just do this inside the MathChat directory:

~$ ruby script/server

Then go to localhost:3000 in your browser (I suggest Google Chrome for better performance) and enjoy MathChat!

MathChat is now hosted at


hg clone math-chat

And, finally a bunch of screenshots!

Regards, Juan Manuel

PHP Web Crawler

I Finished to write a Web Crawler in PHP.

Here’s a Tutorial about it. I hope it will be useful for you.

The web crawler is very easy to use. To run it, just do this in the console:

~$ php main.php

Inside the main directory of the application.

The configuration will be taken from the config.ini file. For example:

host = “localhost:3307”
user = “root”
pass = “root”
db = “jm”

start_url = “”
max_depth = 0
log = “1”

The first 4 parameters are the database connection. I assume this is know to you.

The start_url param is the url to start to craw. Note: The url must be complete! Don’t ignore the http:// or https:// if it correspond.

You can specify the maximum of recursive searches in the  max_depth param. 0 only crawls the start url. 1 crawls the start_url and all the urls inside the given url. 2 All the urls inside the urls given by previous and so on… Warning: A factor of 3 or greater can take for hours, days, month or years!

Finally the log parameter indicates if the application shows the crawled urls in the console.

The config.ini can be edited by the web UI:

It’s very intuitive and you eventually can start to crawl from it. You can also watch all that’s crawled to the moment clicking in the “see what’s crawled” button.

Finally I left a list of features about the PHP Web Crawler:

– The crawler can be run as multiple instances

– It can be run by a cron job.

– All the crawls are saved in a mysql database. It generates the table “urls” to store the crawls.

– For each url it saves the url of source, the url of the destiny and the anchor text.

– Validates the urls via a regular expression. It avoids the links to static data into the site. Including the unnecessary media files. Despite this I can’t ensure that the crawler avoids all the media files. That be more complex to validate.

And here is a demo of 6 processes crawling at the same time.


The crawler is now hosted at:


Regards, Juan Manuel